[ 427.821330] nova-conductor[52103]: Modules with known eventlet monkey patching issues were imported prior to eventlet monkey patching: urllib3. This warning can usually be ignored if the caller is only importing and not executing nova code. [ 429.127104] nova-conductor[52103]: DEBUG oslo_db.sqlalchemy.engines [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52103) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 429.153937] nova-conductor[52103]: DEBUG nova.context [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),74c00a1b-a396-4090-b2c1-a2ad7f5ace70(cell1) {{(pid=52103) load_cells /opt/stack/nova/nova/context.py:464}} [ 429.155801] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52103) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 429.156033] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52103) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 429.156506] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52103) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 429.156851] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52103) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 429.157048] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52103) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 429.158014] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=52103) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 429.163492] nova-conductor[52103]: DEBUG oslo_db.sqlalchemy.engines [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52103) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 429.163872] nova-conductor[52103]: DEBUG oslo_db.sqlalchemy.engines [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52103) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 429.225113] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Acquiring lock "singleton_lock" {{(pid=52103) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:312}} [ 429.225300] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Acquired lock "singleton_lock" {{(pid=52103) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:315}} [ 429.225533] nova-conductor[52103]: DEBUG oslo_concurrency.lockutils [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Releasing lock "singleton_lock" {{(pid=52103) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:333}} [ 429.225970] nova-conductor[52103]: INFO oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Starting 2 workers [ 429.230724] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Started child 52522 {{(pid=52103) _start_child /usr/local/lib/python3.10/dist-packages/oslo_service/service.py:575}} [ 429.236021] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Started child 52523 {{(pid=52103) _start_child /usr/local/lib/python3.10/dist-packages/oslo_service/service.py:575}} [ 429.236137] nova-conductor[52522]: INFO nova.service [-] Starting conductor node (version 0.0.1) [ 429.236347] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Full set of CONF: {{(pid=52103) wait /usr/local/lib/python3.10/dist-packages/oslo_service/service.py:649}} [ 429.236347] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ******************************************************************************** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2589}} [ 429.236347] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] Configuration options gathered from: {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2590}} [ 429.236347] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] command line args: ['--config-file', '/etc/nova/nova.conf'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2591}} [ 429.236347] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] config files: ['/etc/nova/nova.conf'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2592}} [ 429.236512] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ================================================================================ {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2594}} [ 429.236838] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] allow_resize_to_same_host = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.237239] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] arq_binding_timeout = 300 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.237291] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] block_device_allocate_retries = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.237484] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] block_device_allocate_retries_interval = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.237714] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cert = self.pem {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.237898] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute_driver = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.238139] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute_monitors = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.238413] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] config_dir = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.238611] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] config_drive_format = iso9660 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.238741] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] config_file = ['/etc/nova/nova.conf'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.238943] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] config_source = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.239150] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] console_host = devstack {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.239545] nova-conductor[52523]: INFO nova.service [-] Starting conductor node (version 0.0.1) [ 429.239961] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] control_exchange = nova {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.240180] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cpu_allocation_ratio = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.240365] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] daemon = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.240575] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] debug = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.240747] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] default_access_ip_network_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.240925] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] default_availability_zone = nova {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.241088] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] default_ephemeral_format = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.241412] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.241657] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] default_schedule_zone = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.241862] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] disk_allocation_ratio = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.242052] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] enable_new_services = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.242290] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] enabled_apis = ['osapi_compute'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.242492] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] enabled_ssl_apis = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.242686] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] flat_injected = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244042] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] force_config_drive = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244042] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] force_raw_images = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244042] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] graceful_shutdown_timeout = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244042] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] heal_instance_info_cache_interval = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244042] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] host = devstack {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244042] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] initial_cpu_allocation_ratio = 4.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244262] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] initial_disk_allocation_ratio = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244386] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] initial_ram_allocation_ratio = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244993] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.244993] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instance_build_timeout = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.245097] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instance_delete_interval = 300 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.245295] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instance_format = [instance: %(uuid)s] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.245483] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instance_name_template = instance-%08x {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.245661] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instance_usage_audit = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.245850] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instance_usage_audit_period = month {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.246226] nova-conductor[52522]: DEBUG oslo_db.sqlalchemy.engines [None req-4719813c-5ea5-43f7-b6f9-9fb70fc7bb59 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52522) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 429.246394] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.246600] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] instances_path = /opt/stack/data/nova/instances {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.246784] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] internal_service_availability_zone = internal {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.246976] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] key = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.247136] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] live_migration_retry_count = 30 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.247620] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_config_append = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.247620] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.247700] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_dir = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.247871] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.247996] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_options = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.248177] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_rotate_interval = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.248370] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_rotate_interval_type = days {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.248568] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] log_rotation_type = none {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.248700] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.248827] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.248993] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.249186] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.249314] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.249527] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] long_rpc_timeout = 1800 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.249689] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] max_concurrent_builds = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.249838] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] max_concurrent_live_migrations = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.249987] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] max_concurrent_snapshots = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.250147] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] max_local_block_devices = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.250298] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] max_logfile_count = 30 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.250452] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] max_logfile_size_mb = 200 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.250626] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] maximum_instance_delete_attempts = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.250787] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] metadata_listen = 0.0.0.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.251023] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] metadata_listen_port = 8775 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.251192] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] metadata_workers = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.251348] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] migrate_max_retries = -1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.251516] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] mkisofs_cmd = genisoimage {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.251754] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] my_block_storage_ip = 10.180.1.21 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.251880] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] my_ip = 10.180.1.21 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.252045] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] network_allocate_retries = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.252245] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.252722] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] osapi_compute_listen = 0.0.0.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.252722] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] osapi_compute_listen_port = 8774 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.252722] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] osapi_compute_unique_server_name_scope = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.252873] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] osapi_compute_workers = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.253058] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] password_length = 12 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.253227] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] periodic_enable = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.253402] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] periodic_fuzzy_delay = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.253584] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] pointer_model = usbtablet {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.253791] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] preallocate_images = none {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.253948] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] publish_errors = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.254087] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] pybasedir = /opt/stack/nova {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.254260] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ram_allocation_ratio = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.254425] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rate_limit_burst = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.254582] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rate_limit_except_level = CRITICAL {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.254735] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rate_limit_interval = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.254883] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] reboot_timeout = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.255053] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] reclaim_instance_interval = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.255207] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] record = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.255381] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] reimage_timeout_per_gb = 20 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.255531] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] report_interval = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.255851] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rescue_timeout = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.255917] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] reserved_host_cpus = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.256319] nova-conductor[52523]: DEBUG oslo_db.sqlalchemy.engines [None req-1ae56f23-e4be-4794-bd49-de958a46a8ba None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52523) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 429.256573] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] reserved_host_disk_mb = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.256756] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] reserved_host_memory_mb = 512 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.256937] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] reserved_huge_pages = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.257117] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] resize_confirm_window = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.257284] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] resize_fs_using_block_device = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.257439] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] resume_guests_state_on_host_boot = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258303] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258303] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rpc_response_timeout = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258303] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] run_external_periodic_tasks = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258303] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] running_deleted_instance_action = reap {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258303] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] running_deleted_instance_poll_interval = 1800 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258491] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] running_deleted_instance_timeout = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258594] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler_instance_sync_interval = 120 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258765] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_down_time = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.258953] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] servicegroup_driver = db {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] shelved_offload_time = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] shelved_poll_interval = 3600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] shutdown_timeout = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] source_is_ipv6 = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ssl_only = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] state_path = /opt/stack/data/nova {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260355] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] sync_power_state_interval = 600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260355] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] sync_power_state_pool_size = 1000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260448] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] syslog_log_facility = LOG_USER {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260572] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] tempdir = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260739] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] timeout_nbd = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.260895] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] transport_url = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.261074] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] update_resources_interval = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.261228] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] use_cow_images = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.261396] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] use_eventlog = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.261569] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] use_journal = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.261718] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] use_json = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.261872] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] use_rootwrap_daemon = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.262058] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] use_stderr = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.262237] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] use_syslog = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.262408] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vcpu_pin_set = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.262571] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vif_plugging_is_fatal = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.262757] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vif_plugging_timeout = 300 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.262955] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] virt_mkfs = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.263121] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] volume_usage_poll_interval = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.263293] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] watch_log_file = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.263478] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] web = /usr/share/spice-html5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2602}} [ 429.263757] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_concurrency.disable_process_locking = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.263946] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_concurrency.lock_path = /opt/stack/data/nova {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.264164] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.264323] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.264482] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_metrics.metrics_process_name = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.264638] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.264799] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.265037] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.auth_strategy = keystone {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.265226] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.compute_link_prefix = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.265422] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.265609] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.dhcp_domain = novalocal {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.265810] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.enable_instance_password = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.265973] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.glance_link_prefix = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268018] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268018] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.instance_list_cells_batch_strategy = distributed {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268018] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.instance_list_per_project_cells = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268018] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.list_records_by_skipping_down_cells = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268018] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.local_metadata_per_cell = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268018] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.max_limit = 1000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.metadata_cache_expiration = 15 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.neutron_default_tenant_id = default {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.use_forwarded_for = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.use_neutron_default_nets = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.vendordata_dynamic_failure_fatal = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268646] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.vendordata_dynamic_ssl_certfile = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268646] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.vendordata_dynamic_targets = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268646] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.vendordata_jsonfile_path = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.268828] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api.vendordata_providers = ['StaticJSON'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.269092] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.backend = dogpile.cache.memcached {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.269272] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.backend_argument = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.269478] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.config_prefix = cache.oslo {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.269686] nova-conductor[52522]: DEBUG nova.service [None req-4719813c-5ea5-43f7-b6f9-9fb70fc7bb59 None None] Creating RPC server for service conductor {{(pid=52522) start /opt/stack/nova/nova/service.py:182}} [ 429.269836] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.dead_timeout = 60.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.270032] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.debug_cache_backend = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.270218] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.enable_retry_client = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.270394] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.enable_socket_keepalive = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.270560] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.enabled = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.270734] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.expiration_time = 600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.270892] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.hashclient_retry_attempts = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.271523] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.hashclient_retry_delay = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.271523] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_dead_retry = 300 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.271523] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_password = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.271523] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.271741] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.271822] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_pool_maxsize = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.271974] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_pool_unused_timeout = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.272144] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_sasl_enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.272313] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_servers = ['localhost:11211'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.272505] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_socket_timeout = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.272668] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.memcache_username = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.272827] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.proxies = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.272981] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.retry_attempts = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.273151] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.retry_delay = 0.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.273305] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.socket_keepalive_count = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.273457] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.socket_keepalive_idle = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.273614] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.socket_keepalive_interval = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.273761] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.tls_allowed_ciphers = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.273907] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.tls_cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.274089] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.tls_certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.274255] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.tls_enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.274403] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cache.tls_keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.274614] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.274809] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.auth_type = password {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.274980] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.275177] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.catalog_info = volumev3::publicURL {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.275331] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.275514] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.275688] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.cross_az_attach = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.275863] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.debug = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.276037] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.endpoint_template = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.276207] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.http_retries = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.276366] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.276516] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.276696] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.os_region_name = RegionOne {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.276857] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.277016] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cinder.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.277190] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.277359] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.cpu_dedicated_set = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.277551] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.cpu_shared_set = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.277716] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.image_type_exclude_list = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.277875] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.live_migration_wait_for_vif_plug = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.278042] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.max_concurrent_disk_ops = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.278224] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.max_disk_devices_to_attach = -1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.278386] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.278558] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.278712] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.resource_provider_association_refresh = 300 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.278870] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.shutdown_retry_interval = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] conductor.workers = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] console.allowed_origins = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] console.ssl_ciphers = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] console.ssl_minimum_version = default {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] consoleauth.token_ttl = 600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280353] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280353] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280443] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280558] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.connect_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280730] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.connect_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.280882] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.endpoint_override = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.281043] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.281194] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.281343] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.max_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.281508] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.min_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.281662] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.region_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.281822] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.service_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.281971] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.service_type = accelerator {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.282136] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.282287] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.status_code_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.282439] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.status_code_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.282589] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.282764] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.282938] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] cyborg.version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.283127] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.backend = sqlalchemy {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.283368] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.connection = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.283540] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.connection_debug = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.283816] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.connection_parameters = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.283816] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.connection_recycle_time = 3600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.283968] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.connection_trace = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.284417] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.db_inc_retry_interval = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.284417] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.db_max_retries = 20 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.284538] nova-conductor[52523]: DEBUG nova.service [None req-1ae56f23-e4be-4794-bd49-de958a46a8ba None None] Creating RPC server for service conductor {{(pid=52523) start /opt/stack/nova/nova/service.py:182}} [ 429.284670] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.db_max_retry_interval = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.284849] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.db_retry_interval = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.285026] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.max_overflow = 50 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.285186] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.max_pool_size = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.285344] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.max_retries = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.285525] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.mysql_enable_ndb = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.285680] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.mysql_sql_mode = TRADITIONAL {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.285833] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.mysql_wsrep_sync_wait = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.285990] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.pool_timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.286189] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.retry_interval = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.286316] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.slave_connection = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.286483] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.sqlite_synchronous = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.286646] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] database.use_db_reconnect = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.286834] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.backend = sqlalchemy {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.287033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.connection = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.287364] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.connection_debug = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.287364] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.connection_parameters = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.287563] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.connection_recycle_time = 3600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.287730] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.connection_trace = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.287889] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.db_inc_retry_interval = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.288747] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.db_max_retries = 20 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.288747] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.db_max_retry_interval = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.288747] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.db_retry_interval = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.288747] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.max_overflow = 50 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.288747] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.max_pool_size = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.289029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.max_retries = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.289029] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.mysql_enable_ndb = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.289182] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.289322] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.mysql_wsrep_sync_wait = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.289543] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.pool_timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.289709] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.retry_interval = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.289853] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.slave_connection = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.290020] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] api_database.sqlite_synchronous = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.290222] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] devices.enabled_mdev_types = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.290428] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.290598] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ephemeral_storage_encryption.enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.290781] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ephemeral_storage_encryption.key_size = 512 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.291664] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.api_servers = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.291664] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.291664] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.291664] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.291664] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.connect_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.291933] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.connect_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.291933] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.debug = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.292142] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.default_trusted_certificate_ids = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.292311] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.enable_certificate_validation = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293114] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.enable_rbd_download = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293114] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.endpoint_override = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293114] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293114] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293114] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.max_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293401] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.min_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293401] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.num_retries = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293548] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.rbd_ceph_conf = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293843] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.rbd_connect_timeout = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.293843] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.rbd_pool = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.294052] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.rbd_user = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.294384] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.region_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.294384] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.service_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.294620] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.service_type = image {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.294692] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.295079] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.status_code_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.295079] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.status_code_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.295234] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.verify_glance_signatures = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] glance.version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] guestfs.debug = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.config_drive_cdrom = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296308] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.config_drive_inject_password = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296308] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.dynamic_memory_ratio = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296437] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.enable_instance_metrics_collection = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296887] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.enable_remotefx = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.296887] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.instances_path_share = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.297031] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.iscsi_initiator_list = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.297080] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.limit_cpu_features = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.297219] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.mounted_disk_query_retry_count = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.297405] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.mounted_disk_query_retry_interval = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.297587] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.power_state_check_timeframe = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298024] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.power_state_event_polling_interval = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298024] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.qemu_img_cmd = qemu-img.exe {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298158] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.use_multipath_io = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298240] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.volume_attach_retry_count = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298393] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.volume_attach_retry_interval = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298548] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.vswitch_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298701] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] hyperv.wait_soft_reboot_seconds = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.298880] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] mks.enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.299557] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.299737] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] image_cache.manager_interval = 2400 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.300034] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] image_cache.precache_concurrency = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.300111] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] image_cache.remove_unused_base_images = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.300241] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.300403] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301440] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] image_cache.subdirectory_name = _base {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301440] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.api_max_retries = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301440] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.api_retry_interval = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301440] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301440] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.auth_type = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301723] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301723] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.301916] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.302018] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.connect_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.302171] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.connect_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.302622] nova-conductor[52522]: DEBUG nova.service [None req-4719813c-5ea5-43f7-b6f9-9fb70fc7bb59 None None] Join ServiceGroup membership for this service conductor {{(pid=52522) start /opt/stack/nova/nova/service.py:199}} [ 429.302781] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.endpoint_override = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.302956] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.303399] nova-conductor[52522]: DEBUG nova.servicegroup.drivers.db [None req-4719813c-5ea5-43f7-b6f9-9fb70fc7bb59 None None] DB_Driver: join new ServiceGroup member devstack to the conductor group, service = {{(pid=52522) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 429.303482] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.303482] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.max_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304188] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.min_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304188] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.partition_key = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304188] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.peer_list = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304188] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.region_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304418] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.serial_console_state_timeout = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304418] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.service_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304604] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.service_type = baremetal {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304762] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.304912] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.status_code_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.305076] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.status_code_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.305327] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.305397] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.305562] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ironic.version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.305767] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.305957] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] key_manager.fixed_key = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.306602] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.306728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.barbican_api_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.306895] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.barbican_endpoint = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.307133] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.barbican_endpoint_type = public {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.307269] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.barbican_region_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.307489] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.307632] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.307804] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.307945] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.308113] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.308292] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.number_of_retries = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.308431] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.retry_delay = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.308663] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.send_service_user_token = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.308764] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.309235] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.309235] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.verify_ssl = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.309235] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican.verify_ssl_path = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.309415] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.309570] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.auth_type = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.309726] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.309883] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.310728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.310728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.310728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.310728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.310728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] barbican_service_user.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311019] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.approle_role_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311081] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.approle_secret_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311192] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311339] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311500] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311669] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311803] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.311986] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.kv_mountpoint = secret {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.312168] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.kv_version = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.312380] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.namespace = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.312487] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.root_token_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.312651] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.312794] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.ssl_ca_crt_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.312962] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.313205] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.use_ssl = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.313781] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.313781] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.313781] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.314234] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.314234] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.connect_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.314234] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.connect_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.314361] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.endpoint_override = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.314436] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.314777] nova-conductor[52523]: DEBUG nova.service [None req-1ae56f23-e4be-4794-bd49-de958a46a8ba None None] Join ServiceGroup membership for this service conductor {{(pid=52523) start /opt/stack/nova/nova/service.py:199}} [ 429.314923] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.315258] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.max_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.315331] nova-conductor[52523]: DEBUG nova.servicegroup.drivers.db [None req-1ae56f23-e4be-4794-bd49-de958a46a8ba None None] DB_Driver: join new ServiceGroup member devstack to the conductor group, service = {{(pid=52523) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 429.315431] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.min_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.315604] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.region_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.315875] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.service_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.315946] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.service_type = identity {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.316102] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.316264] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.status_code_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.316422] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.status_code_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.316613] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.316802] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.316960] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] keystone.version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.317221] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.connection_uri = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.317420] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.cpu_mode = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.317606] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.cpu_model_extra_flags = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.317777] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.cpu_models = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.317945] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.cpu_power_governor_high = performance {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.318125] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.cpu_power_governor_low = powersave {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.318288] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.cpu_power_management = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.318480] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.318663] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.device_detach_attempts = 8 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.318841] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.device_detach_timeout = 20 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.319012] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.disk_cachemodes = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.319182] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.disk_prefix = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.319343] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.enabled_perf_events = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.319546] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.file_backed_memory = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.319711] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.gid_maps = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.319890] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.hw_disk_discard = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.320058] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.hw_machine_type = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.320249] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.images_rbd_ceph_conf = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.320429] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.320619] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.320780] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.images_rbd_glance_store_name = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.320950] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.images_rbd_pool = rbd {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.321126] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.images_type = default {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.321283] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.images_volume_group = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.321442] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.inject_key = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.321603] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.inject_partition = -2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.321762] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.inject_password = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.321946] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.iscsi_iface = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.322119] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.iser_use_multipath = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.322282] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_bandwidth = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.322441] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_completion_timeout = 800 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.322600] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_downtime = 500 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.322755] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_downtime_delay = 75 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.322912] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_downtime_steps = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.323085] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_inbound_addr = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.323248] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_permit_auto_converge = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.323414] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_permit_post_copy = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.323569] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_scheme = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.323735] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_timeout_action = abort {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.323892] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_tunnelled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.324057] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_uri = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.324219] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.live_migration_with_native_tls = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.324397] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.max_queues = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.324561] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.mem_stats_period_seconds = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.324750] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.nfs_mount_options = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.325140] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.nfs_mount_point_base = /opt/stack/data/nova/mnt {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.325327] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.num_aoe_discover_tries = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.325496] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.num_iser_scan_tries = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.325653] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.num_memory_encrypted_guests = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.325812] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.num_nvme_discover_tries = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.325971] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.num_pcie_ports = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.326150] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.num_volume_scan_tries = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.326367] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.pmem_namespaces = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.326531] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.quobyte_client_cfg = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.326766] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.quobyte_mount_point_base = /opt/stack/data/nova/mnt {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.326931] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rbd_connect_timeout = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.327108] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.327271] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.327460] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rbd_secret_uuid = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.327641] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rbd_user = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.327805] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.realtime_scheduler_priority = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.327996] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.remote_filesystem_transport = ssh {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.328171] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rescue_image_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.328323] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rescue_kernel_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.328477] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rescue_ramdisk_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.328646] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rng_dev_path = /dev/urandom {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.328800] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.rx_queue_size = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.328966] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.smbfs_mount_options = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.329211] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.smbfs_mount_point_base = /opt/stack/data/nova/mnt {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.329457] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.snapshot_compression = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.329637] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.snapshot_image_format = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.329857] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.330033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.sparse_logical_volumes = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.330203] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.swtpm_enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.330373] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.swtpm_group = tss {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.330546] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.swtpm_user = tss {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.330708] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.sysinfo_serial = unique {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.330864] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.tx_queue_size = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.331032] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.uid_maps = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.331200] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.use_virtio_for_bridges = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.331366] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.virt_type = kvm {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.331534] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.volume_clear = zero {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.331701] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.volume_clear_size = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.331880] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.volume_use_multipath = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.332066] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.vzstorage_cache_path = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.332237] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.332400] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.vzstorage_mount_group = qemu {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.332562] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.vzstorage_mount_opts = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.332722] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.332937] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/nova/mnt {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.333116] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.vzstorage_mount_user = stack {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.333276] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.333475] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.333653] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.auth_type = password {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.333817] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.333964] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.334136] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.334293] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.connect_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.334449] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.connect_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.334614] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.default_floating_pool = public {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.334768] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.endpoint_override = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.334932] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.extension_sync_interval = 600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.335098] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.http_retries = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.335258] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.335412] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.335580] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.max_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.335737] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.metadata_proxy_shared_secret = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.335891] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.min_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.336093] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.ovs_bridge = br-int {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.336261] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.physnets = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.336444] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.region_name = RegionOne {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.336621] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.service_metadata_proxy = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.336784] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.service_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.336981] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.service_type = network {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.337165] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.337325] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.status_code_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.337481] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.status_code_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.337636] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.337812] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.337968] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] neutron.version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.338177] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] notifications.bdms_in_notifications = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.338353] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] notifications.default_level = INFO {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.338543] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] notifications.notification_format = unversioned {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.338706] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] notifications.notify_on_state_change = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.338879] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.339061] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] pci.alias = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.339227] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] pci.device_spec = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.339418] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] pci.report_in_placement = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.339613] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.339792] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.auth_type = password {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.339981] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.auth_url = http://10.180.1.21/identity {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.340156] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.340313] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.340518] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.340686] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.connect_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.340843] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.connect_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.340998] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.default_domain_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.341171] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.default_domain_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.341325] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.domain_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.341475] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.domain_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.341629] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.endpoint_override = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.341784] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.341935] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.342097] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.max_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.342252] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.min_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.342412] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.password = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.342575] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.project_domain_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.342728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.project_domain_name = Default {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.342887] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.project_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.343071] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.project_name = service {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.343239] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.region_name = RegionOne {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.343479] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.service_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.343550] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.service_type = placement {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.343758] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.343932] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.status_code_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.344100] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.status_code_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.344258] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.system_scope = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.344410] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.344563] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.trust_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.344760] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.user_domain_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.344884] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.user_domain_name = Default {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.345049] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.user_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.345222] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.username = placement {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.345396] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.345554] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] placement.version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.345753] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.cores = 20 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.345999] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.count_usage_from_placement = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.346209] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.346405] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.injected_file_content_bytes = 10240 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.346573] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.injected_file_path_length = 255 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.346736] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.injected_files = 5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.346898] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.instances = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.347069] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.key_pairs = 100 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.347234] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.metadata_items = 128 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.347400] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.ram = 51200 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.347561] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.recheck_quota = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.347722] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.server_group_members = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.347880] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] quota.server_groups = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.348055] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rdp.enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.348370] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] rdp.html5_proxy_base_url = http://127.0.0.1:6083/ {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.348583] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.348769] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.348960] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.image_metadata_prefilter = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.349164] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.349329] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.max_attempts = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.349517] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.max_placement_results = 1000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.349711] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.349874] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.query_placement_for_availability_zone = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.350064] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.query_placement_for_image_type_support = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.350247] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.350460] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] scheduler.workers = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.350670] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.350843] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.351062] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.351238] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.351400] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.351563] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.351719] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.351930] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.352111] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.host_subset_size = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.352291] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.image_properties_default_architecture = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.352456] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.352619] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.isolated_hosts = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.352800] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.isolated_images = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.352961] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.max_instances_per_host = 50 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.353132] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.353298] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.pci_in_placement = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.353452] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.353611] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.353770] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.353926] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.354092] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.354252] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.354440] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.track_instance_changes = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.354617] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.354788] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] metrics.required = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.354948] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] metrics.weight_multiplier = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.355117] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] metrics.weight_of_unavailable = -10000.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.355305] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] metrics.weight_setting = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.355613] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.355786] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] serial_console.enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.355979] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] serial_console.port_range = 10000:20000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.356162] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.356348] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.356513] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] serial_console.serialproxy_port = 6083 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.356677] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.356846] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.auth_type = password {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.357033] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.357174] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.357341] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.357523] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.357686] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.357861] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.send_service_user_token = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.358031] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.358194] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] service_user.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.358360] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.agent_enabled = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.358550] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.358908] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.359150] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.html5proxy_host = 0.0.0.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.359345] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.html5proxy_port = 6082 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.359534] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.image_compression = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.359693] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.jpeg_compression = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.359850] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.playback_compression = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.360050] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.server_listen = 127.0.0.1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.360244] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.360405] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.streaming_mode = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.360567] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] spice.zlib_compression = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.360727] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] upgrade_levels.baseapi = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.360893] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] upgrade_levels.cert = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.361096] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] upgrade_levels.compute = auto {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.361258] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] upgrade_levels.conductor = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.361412] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] upgrade_levels.scheduler = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.361578] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.361746] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.auth_type = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.361894] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.362048] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.362211] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.362366] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.362517] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.362683] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.362836] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vendordata_dynamic_auth.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.363049] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.api_retry_count = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.363214] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.ca_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.363369] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.cache_prefix = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.363521] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.cluster_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.363681] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.connection_pool_size = 10 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.363834] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.console_delay_seconds = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.363986] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.datastore_regex = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.364153] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.host_ip = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.364307] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.host_password = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.364466] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.host_port = 443 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.364619] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.host_username = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.364773] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.364935] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.integration_bridge = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.365105] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.maximum_objects = 100 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.365260] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.pbm_default_policy = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.365415] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.pbm_enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.365568] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.pbm_wsdl_location = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.365728] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.365880] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.serial_port_proxy_uri = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.366039] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.serial_port_service_uri = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.366205] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.task_poll_interval = 0.5 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.366363] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.use_linked_clone = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.366525] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.vnc_keymap = en-us {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.366689] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.vnc_port = 5900 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.366843] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vmware.vnc_port_total = 10000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.367054] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.auth_schemes = ['none'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.367223] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.enabled = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.367568] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.367758] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.367928] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.novncproxy_port = 6080 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.368120] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.server_listen = 127.0.0.1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.368292] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.368454] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.vencrypt_ca_certs = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.368608] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.vencrypt_client_cert = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.368760] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] vnc.vencrypt_client_key = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.368970] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.369147] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.disable_deep_image_inspection = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.369307] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.disable_fallback_pcpu_query = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.369496] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.disable_group_policy_check_upcall = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.369661] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.369820] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.disable_rootwrap = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.369978] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.enable_numa_live_migration = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.370150] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.370311] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.370515] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.handle_virt_lifecycle_events = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.370661] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.libvirt_disable_apic = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.370825] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.never_download_image_if_on_rbd = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.370987] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.371161] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.371320] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.371477] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.371640] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.371804] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.371959] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.372128] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.372291] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.372494] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.372663] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.client_socket_timeout = 900 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.372823] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.default_pool_size = 1000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.373026] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.keep_alive = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.373275] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.max_header_line = 16384 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.373365] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.secure_proxy_ssl_header = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.373523] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.ssl_ca_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.373681] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.ssl_cert_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.373838] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.ssl_key_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.374006] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.tcp_keepidle = 600 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.374187] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.374356] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] zvm.ca_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.374507] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] zvm.cloud_connector_url = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.374731] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] zvm.image_tmp_path = /opt/stack/data/nova/images {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.374891] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] zvm.reachable_timeout = 300 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.375130] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.enforce_new_defaults = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.375319] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.enforce_scope = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.375508] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.policy_default_rule = default {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.375711] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.375902] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.policy_file = policy.yaml {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.376110] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.376292] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.376449] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.376624] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.376801] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.376998] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.377209] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.377437] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.connection_string = messaging:// {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.377634] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.enabled = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.377822] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.es_doc_type = notification {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.378006] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.es_scroll_size = 10000 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.378179] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.es_scroll_time = 2m {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.378341] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.filter_error_trace = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.378505] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.hmac_keys = SECRET_KEY {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.378667] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.sentinel_service_name = mymaster {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.378853] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.socket_timeout = 0.1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.379030] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] profiler.trace_sqlalchemy = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.379236] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] remote_debug.host = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.379438] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] remote_debug.port = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.379639] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.379803] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.379969] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.380144] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.380312] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.380497] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.380674] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.380835] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.heartbeat_rate = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.380999] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.381171] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.381342] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.381505] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.381676] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.381840] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.382007] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.382189] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.382350] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.382509] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.382672] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.382830] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.382988] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.383161] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.383317] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.383477] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.383638] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.383805] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.ssl = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.383970] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.384148] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.384308] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.384473] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.384638] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_rabbit.ssl_version = {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.384853] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.385056] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_notifications.retry = -1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.385257] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.385429] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_messaging_notifications.transport_url = **** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.385640] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.auth_section = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.385805] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.auth_type = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.385962] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.cafile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.386128] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.certfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.386311] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.collect_timing = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.386470] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.connect_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.386628] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.connect_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.386800] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.endpoint_id = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.386955] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.endpoint_override = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.387125] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.insecure = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.387282] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.keyfile = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.387448] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.max_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.387613] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.min_version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.387771] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.region_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.387923] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.service_name = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.388084] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.service_type = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.388246] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.split_loggers = False {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.388400] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.status_code_retries = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.388560] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.status_code_retry_delay = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.388713] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.timeout = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.388866] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.valid_interfaces = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.389032] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_limit.version = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.389239] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_reports.file_event_handler = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.389460] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_reports.file_event_handler_interval = 1 {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.389590] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] oslo_reports.log_dir = None {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2609}} [ 429.389722] nova-conductor[52103]: DEBUG oslo_service.service [None req-fd617185-b684-42e2-9af9-e9bae0e9591a None None] ******************************************************************************** {{(pid=52103) log_opt_values /usr/local/lib/python3.10/dist-packages/oslo_config/cfg.py:2613}} [ 519.055404] nova-conductor[52523]: DEBUG oslo_db.sqlalchemy.engines [None req-c8d32dc3-a4fe-445a-b5a2-f857248bb81b None None] Parent process 52103 forked (52523) with an open database connection, which is being discarded and recreated. {{(pid=52523) checkout /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:434}} [ 554.317561] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Took 0.57 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 554.369051] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.369407] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.370938] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.381416] nova-conductor[52523]: DEBUG oslo_db.sqlalchemy.engines [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52523) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 554.466828] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.466828] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.468039] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.468039] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.468039] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.468039] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.480448] nova-conductor[52523]: DEBUG oslo_db.sqlalchemy.engines [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52523) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 554.504192] nova-conductor[52523]: DEBUG nova.quota [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Getting quotas for project b9d958b398104133b85a3f555684d617. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 554.507612] nova-conductor[52523]: DEBUG nova.quota [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Getting quotas for user ff3acb692cd64137b803aad2fd0b6822 and project b9d958b398104133b85a3f555684d617. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 554.518035] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] [instance: 15e44d1f-ae9b-4ff7-841c-90acc81cf38b] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 554.518698] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.518903] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.519078] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.528197] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] [instance: 15e44d1f-ae9b-4ff7-841c-90acc81cf38b] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 554.528197] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.528197] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.528337] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.564295] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.564646] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.564832] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.565131] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=52523) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:312}} [ 554.565276] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Acquired lock "compute-rpcapi-router" {{(pid=52523) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:315}} [ 554.565810] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.565995] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.566172] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.566506] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 554.566679] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 554.566831] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 554.574401] nova-conductor[52523]: INFO nova.compute.rpcapi [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Automatically selected compute RPC version 6.2 from minimum service version 66 [ 554.574401] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-8c1101f0-a10e-4a2e-a3cb-ae83c046c9ec None None] Releasing lock "compute-rpcapi-router" {{(pid=52523) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:333}} [ 560.810078] nova-conductor[52522]: DEBUG oslo_db.sqlalchemy.engines [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Parent process 52103 forked (52522) with an open database connection, which is being discarded and recreated. {{(pid=52522) checkout /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:434}} [ 561.025875] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Took 0.21 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 561.056739] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.057124] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.058809] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.002s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.066429] nova-conductor[52522]: DEBUG oslo_db.sqlalchemy.engines [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52522) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 561.130127] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.130399] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.130915] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.131545] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.131874] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.132137] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.140313] nova-conductor[52522]: DEBUG oslo_db.sqlalchemy.engines [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=52522) _check_effective_sql_mode /usr/local/lib/python3.10/dist-packages/oslo_db/sqlalchemy/engines.py:335}} [ 561.155082] nova-conductor[52522]: DEBUG nova.quota [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Getting quotas for project 4b3edaaa3fdc4f73b49b8e57e04b8fa0. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 561.159756] nova-conductor[52522]: DEBUG nova.quota [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Getting quotas for user 5208b8d162bc46d489a34997aaebbaa2 and project 4b3edaaa3fdc4f73b49b8e57e04b8fa0. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 561.167485] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 8a0d9829-6759-4593-9230-459a546a5908] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 561.167485] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.167485] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.167485] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.173074] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 8a0d9829-6759-4593-9230-459a546a5908] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 561.173745] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.173878] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.174060] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.208779] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.209038] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.209220] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.209583] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=52522) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:312}} [ 561.209726] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquired lock "compute-rpcapi-router" {{(pid=52522) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:315}} [ 561.210454] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.210535] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.210654] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.211038] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 561.211224] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 561.211380] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 561.223194] nova-conductor[52522]: INFO nova.compute.rpcapi [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Automatically selected compute RPC version 6.2 from minimum service version 66 [ 561.223714] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b7b23378-b9ac-4a9f-8c19-45c7272ae937 None None] Releasing lock "compute-rpcapi-router" {{(pid=52522) lock /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:333}} [ 566.297405] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Took 0.21 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 566.320395] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 566.320395] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 566.320395] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 566.369957] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 566.370216] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 566.370685] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 566.371142] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 566.371142] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 566.372372] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 566.380785] nova-conductor[52523]: DEBUG nova.quota [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Getting quotas for project 8daf8ee8cb0542f099584ca77665f732. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 566.383912] nova-conductor[52523]: DEBUG nova.quota [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Getting quotas for user 670367f892a9462f982808a51b5d890b and project 8daf8ee8cb0542f099584ca77665f732. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 566.390373] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] [instance: c4400e80-4457-4a8a-8588-f594e5993cde] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 566.393201] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 566.393201] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 566.393201] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 566.397311] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] [instance: c4400e80-4457-4a8a-8588-f594e5993cde] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 566.397311] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 566.397311] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 566.397809] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 566.420511] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 566.420511] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 566.420511] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 568.776337] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Took 0.14 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 568.792699] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 568.792955] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 568.793305] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 568.834965] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 568.835194] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 568.835736] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 568.835736] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 568.835960] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 568.836066] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 568.848086] nova-conductor[52523]: DEBUG nova.quota [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Getting quotas for project 7864d1883b764303a345385abda4d8ae. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 568.851709] nova-conductor[52523]: DEBUG nova.quota [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Getting quotas for user 78fa507f736a47b49c3c81224a423458 and project 7864d1883b764303a345385abda4d8ae. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 568.859030] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] [instance: 68c87b51-b90a-47cc-bec1-05f7c389fc14] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 568.860599] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 568.860599] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 568.860599] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 568.863350] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] [instance: 68c87b51-b90a-47cc-bec1-05f7c389fc14] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 568.864317] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 568.864389] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 568.864972] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 568.882663] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 568.882895] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 568.883076] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 569.015411] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Took 0.19 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 569.030795] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 569.030795] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 569.031027] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 569.090690] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 569.090968] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 569.091164] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 569.091562] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 569.091731] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 569.091893] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 569.111498] nova-conductor[52522]: DEBUG nova.quota [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Getting quotas for project f0031e355e57421a8d48003a7eb717db. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 569.116083] nova-conductor[52522]: DEBUG nova.quota [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Getting quotas for user c06f3b2e0bd4459696b6724fa90f3809 and project f0031e355e57421a8d48003a7eb717db. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 569.129320] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: 7a4778b7-5ffc-4641-b968-d0304fd67ee0] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 569.129320] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 569.129320] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 569.129320] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 569.133140] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: 7a4778b7-5ffc-4641-b968-d0304fd67ee0] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 569.137705] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 569.137705] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 569.137705] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 569.152252] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 569.152252] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 569.152252] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 572.452141] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Took 0.15 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 572.466577] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 572.467050] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 572.467337] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 572.499599] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 572.499822] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 572.500074] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 572.500350] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 572.500562] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 572.500761] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 572.517650] nova-conductor[52522]: DEBUG nova.quota [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Getting quotas for project f500e2d8bd8b4db28dc4c1f088d12990. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 572.521780] nova-conductor[52522]: DEBUG nova.quota [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Getting quotas for user 7a11dbc35f434cfab97abf7033b16758 and project f500e2d8bd8b4db28dc4c1f088d12990. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 572.528120] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 37916d26-1b5e-4991-83a2-ca5a5b00c2ac] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 572.529009] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 572.529112] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 572.529313] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 572.532771] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 37916d26-1b5e-4991-83a2-ca5a5b00c2ac] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 572.534381] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 572.534381] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 572.534381] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 572.549790] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 572.550062] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 572.550191] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 573.776803] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Took 0.17 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 573.796397] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 573.796659] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 573.796877] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 573.847781] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 573.847781] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 573.847781] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 573.847781] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 573.848167] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 573.848167] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 573.856540] nova-conductor[52523]: DEBUG nova.quota [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Getting quotas for project 216628e9ba21424994765f97d38f1fcc. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 573.859242] nova-conductor[52523]: DEBUG nova.quota [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Getting quotas for user 61ceb0491981460e907a6e5e77cb19ed and project 216628e9ba21424994765f97d38f1fcc. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 573.866474] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] [instance: 363f5261-d589-4f99-b7dd-ab8f16cefee3] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 573.867280] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 573.867626] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 573.867911] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 573.874779] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] [instance: 363f5261-d589-4f99-b7dd-ab8f16cefee3] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 573.874779] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 573.874779] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 573.875133] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 573.904019] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 573.904019] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 573.904019] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 576.089308] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Took 0.24 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 576.110983] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 576.110983] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 576.110983] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 576.160127] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 576.160951] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 576.160951] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 576.161158] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 576.161392] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 576.161553] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 576.177384] nova-conductor[52522]: DEBUG nova.quota [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Getting quotas for project 730d218125d1484687ab1b68a1e73d2e. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 576.178051] nova-conductor[52522]: DEBUG nova.quota [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Getting quotas for user f71b64aa4f71462bb5f6eb5b4083ee63 and project 730d218125d1484687ab1b68a1e73d2e. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 576.188086] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: ce222a29-3611-45b3-9664-87ae2fb1b1b8] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 576.188086] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 576.188086] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 576.188086] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 576.192366] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: ce222a29-3611-45b3-9664-87ae2fb1b1b8] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 576.194269] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 576.194924] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 576.195293] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 576.218248] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 576.218248] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 576.218248] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 578.724088] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Took 0.14 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 578.736986] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 578.737317] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 578.737416] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 578.766110] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 578.766307] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 578.766483] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 578.766891] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 578.767037] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 578.767226] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 578.775977] nova-conductor[52522]: DEBUG nova.quota [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Getting quotas for project dc17a2829e0b4b7298ae746628595053. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 578.778513] nova-conductor[52522]: DEBUG nova.quota [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Getting quotas for user f525273b7a5e4b9d836ee998eefc8ee2 and project dc17a2829e0b4b7298ae746628595053. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 578.784558] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] [instance: 9214a18f-c22d-4e24-980e-7241a2b993bd] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 578.785098] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 578.785327] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 578.785486] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 578.789051] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] [instance: 9214a18f-c22d-4e24-980e-7241a2b993bd] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 578.791518] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 578.791518] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 578.791518] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 578.804714] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 578.804955] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 578.805148] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 581.581105] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Took 0.15 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 581.605533] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 581.605673] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 581.605852] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 581.646528] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 581.646987] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 581.647140] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 581.647513] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 581.647743] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 581.647858] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 581.662911] nova-conductor[52522]: DEBUG nova.quota [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Getting quotas for project 730d218125d1484687ab1b68a1e73d2e. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 581.666733] nova-conductor[52522]: DEBUG nova.quota [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Getting quotas for user f71b64aa4f71462bb5f6eb5b4083ee63 and project 730d218125d1484687ab1b68a1e73d2e. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 581.684063] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: 3ab1b905-cd6f-4d2b-a244-f85e56f796d3] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 581.684622] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 581.684842] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 581.685024] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 581.689038] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: 3ab1b905-cd6f-4d2b-a244-f85e56f796d3] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 581.691038] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 581.691038] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 581.691038] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 581.712569] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 581.715297] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 581.715506] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.003s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 582.925299] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Took 0.16 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 582.940713] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 582.940713] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 582.940903] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 582.984549] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 582.984705] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 582.984880] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 582.985204] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 582.986027] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 582.986027] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 582.999115] nova-conductor[52522]: DEBUG nova.quota [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Getting quotas for project 6bbeaa0167e147da8a6002a03f3ed43d. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 583.002359] nova-conductor[52522]: DEBUG nova.quota [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Getting quotas for user aee8538ca1c447babd57e5ee96214faf and project 6bbeaa0167e147da8a6002a03f3ed43d. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 583.013123] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] [instance: 036d6de2-f69b-4714-b89e-9c4307253675] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 583.013349] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.013567] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.013742] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.016758] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] [instance: 036d6de2-f69b-4714-b89e-9c4307253675] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 583.018030] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.018030] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.018030] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.040485] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.040485] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.040485] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.637168] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Took 0.16 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 583.652803] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.653143] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.653372] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.687638] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.687638] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.687638] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.687638] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.687927] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.687927] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.699833] nova-conductor[52522]: DEBUG nova.quota [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Getting quotas for project 215efdf54af34adc8db180a90116d8bd. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 583.705728] nova-conductor[52522]: DEBUG nova.quota [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Getting quotas for user f9501d1d780a405a8c0ba4b0e90a054c and project 215efdf54af34adc8db180a90116d8bd. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 583.713263] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] [instance: 5888cc9f-7341-4f9c-a93c-dd5ec95f7369] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 583.713852] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.714101] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.714280] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.720398] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] [instance: 5888cc9f-7341-4f9c-a93c-dd5ec95f7369] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 583.721223] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.721438] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.721606] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 583.744386] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 583.744620] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 583.744797] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 587.986635] nova-conductor[52522]: Traceback (most recent call last): [ 587.986635] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 587.986635] nova-conductor[52522]: return func(*args, **kwargs) [ 587.986635] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 587.986635] nova-conductor[52522]: selections = self._select_destinations( [ 587.986635] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 587.986635] nova-conductor[52522]: selections = self._schedule( [ 587.986635] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 587.986635] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 587.986635] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 587.986635] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 587.986635] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 587.986635] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 587.987971] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.989442] nova-conductor[52522]: ERROR nova.conductor.manager [ 587.997919] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 587.997919] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 587.998339] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 588.079945] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] [instance: 0255d64b-6186-4195-89f6-d8ed5c6539df] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 588.080812] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 588.081120] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 588.081268] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 588.087817] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 588.087817] nova-conductor[52522]: Traceback (most recent call last): [ 588.087817] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 588.087817] nova-conductor[52522]: return func(*args, **kwargs) [ 588.087817] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 588.087817] nova-conductor[52522]: selections = self._select_destinations( [ 588.087817] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 588.087817] nova-conductor[52522]: selections = self._schedule( [ 588.087817] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 588.087817] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 588.087817] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 588.087817] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 588.087817] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 588.087817] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 588.089178] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-11e94950-5185-4ff3-90bc-6a753cd28b26 tempest-FloatingIPsAssociationTestJSON-160743284 tempest-FloatingIPsAssociationTestJSON-160743284-project-member] [instance: 0255d64b-6186-4195-89f6-d8ed5c6539df] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 590.572926] nova-conductor[52522]: Traceback (most recent call last): [ 590.572926] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 590.572926] nova-conductor[52522]: return func(*args, **kwargs) [ 590.572926] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 590.572926] nova-conductor[52522]: selections = self._select_destinations( [ 590.572926] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 590.572926] nova-conductor[52522]: selections = self._schedule( [ 590.572926] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 590.572926] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 590.572926] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 590.572926] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 590.572926] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 590.572926] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 590.574087] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.574868] nova-conductor[52522]: ERROR nova.conductor.manager [ 590.585335] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 590.585335] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 590.585335] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 590.642646] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] [instance: f949e50d-1d1c-4dcd-9423-0ac9e1cc276d] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 590.643407] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 590.643663] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 590.643850] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 590.647688] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 590.647688] nova-conductor[52522]: Traceback (most recent call last): [ 590.647688] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 590.647688] nova-conductor[52522]: return func(*args, **kwargs) [ 590.647688] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 590.647688] nova-conductor[52522]: selections = self._select_destinations( [ 590.647688] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 590.647688] nova-conductor[52522]: selections = self._schedule( [ 590.647688] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 590.647688] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 590.647688] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 590.647688] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 590.647688] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 590.647688] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 590.648272] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-beb04fc9-dddd-4d93-9faa-0a111fc9788f tempest-InstanceActionsTestJSON-1536757875 tempest-InstanceActionsTestJSON-1536757875-project-member] [instance: f949e50d-1d1c-4dcd-9423-0ac9e1cc276d] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.157130] nova-conductor[52523]: Traceback (most recent call last): [ 591.157130] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 591.157130] nova-conductor[52523]: return func(*args, **kwargs) [ 591.157130] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 591.157130] nova-conductor[52523]: selections = self._select_destinations( [ 591.157130] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 591.157130] nova-conductor[52523]: selections = self._schedule( [ 591.157130] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 591.157130] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 591.157130] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 591.157130] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 591.157130] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 591.157130] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 591.158319] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.159152] nova-conductor[52523]: ERROR nova.conductor.manager [ 591.164890] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 591.165140] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 591.165320] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 591.225278] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] [instance: 37183923-51df-4845-9dc4-5b5bafa1740d] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 591.226026] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 591.226333] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 591.226601] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 591.236029] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 591.236029] nova-conductor[52523]: Traceback (most recent call last): [ 591.236029] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 591.236029] nova-conductor[52523]: return func(*args, **kwargs) [ 591.236029] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 591.236029] nova-conductor[52523]: selections = self._select_destinations( [ 591.236029] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 591.236029] nova-conductor[52523]: selections = self._schedule( [ 591.236029] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 591.236029] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 591.236029] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 591.236029] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 591.236029] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 591.236029] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.236917] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-a135a338-493a-4fd6-8ee9-ff120735aa7c tempest-ImagesNegativeTestJSON-1784347717 tempest-ImagesNegativeTestJSON-1784347717-project-member] [instance: 37183923-51df-4845-9dc4-5b5bafa1740d] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.654282] nova-conductor[52522]: Traceback (most recent call last): [ 591.654282] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 591.654282] nova-conductor[52522]: return func(*args, **kwargs) [ 591.654282] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 591.654282] nova-conductor[52522]: selections = self._select_destinations( [ 591.654282] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 591.654282] nova-conductor[52522]: selections = self._schedule( [ 591.654282] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 591.654282] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 591.654282] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 591.654282] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 591.654282] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 591.654282] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 591.655149] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.655760] nova-conductor[52522]: ERROR nova.conductor.manager [ 591.661286] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 591.661519] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 591.661693] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 591.707236] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] [instance: df66228e-5eea-4fc3-b5c8-502103177cb6] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 591.707236] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 591.707236] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 591.707477] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 591.710305] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 591.710305] nova-conductor[52522]: Traceback (most recent call last): [ 591.710305] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 591.710305] nova-conductor[52522]: return func(*args, **kwargs) [ 591.710305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 591.710305] nova-conductor[52522]: selections = self._select_destinations( [ 591.710305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 591.710305] nova-conductor[52522]: selections = self._schedule( [ 591.710305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 591.710305] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 591.710305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 591.710305] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 591.710305] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 591.710305] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 591.710902] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-21c21983-0a04-48fc-bc2d-576761ff67ae tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] [instance: df66228e-5eea-4fc3-b5c8-502103177cb6] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.082515] nova-conductor[52523]: Traceback (most recent call last): [ 592.082515] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 592.082515] nova-conductor[52523]: return func(*args, **kwargs) [ 592.082515] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 592.082515] nova-conductor[52523]: selections = self._select_destinations( [ 592.082515] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 592.082515] nova-conductor[52523]: selections = self._schedule( [ 592.082515] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 592.082515] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 592.082515] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 592.082515] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 592.082515] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 592.082515] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 592.083144] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.083745] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.091729] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 592.091729] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 592.091863] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 592.137150] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] [instance: 97b28081-6703-4ce3-8300-e196cb6abe49] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 592.137889] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 592.138119] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 592.138295] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 592.141346] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 592.141346] nova-conductor[52523]: Traceback (most recent call last): [ 592.141346] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 592.141346] nova-conductor[52523]: return func(*args, **kwargs) [ 592.141346] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 592.141346] nova-conductor[52523]: selections = self._select_destinations( [ 592.141346] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 592.141346] nova-conductor[52523]: selections = self._schedule( [ 592.141346] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 592.141346] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 592.141346] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 592.141346] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 592.141346] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 592.141346] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.141987] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-aa497205-1de8-4840-9d6b-92d19293d846 tempest-ImagesOneServerTestJSON-1288149506 tempest-ImagesOneServerTestJSON-1288149506-project-member] [instance: 97b28081-6703-4ce3-8300-e196cb6abe49] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.710552] nova-conductor[52523]: Traceback (most recent call last): [ 592.710552] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 592.710552] nova-conductor[52523]: return func(*args, **kwargs) [ 592.710552] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 592.710552] nova-conductor[52523]: selections = self._select_destinations( [ 592.710552] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 592.710552] nova-conductor[52523]: selections = self._schedule( [ 592.710552] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 592.710552] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 592.710552] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 592.710552] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 592.710552] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 592.710552] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 592.711703] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.712511] nova-conductor[52523]: ERROR nova.conductor.manager [ 592.717267] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 592.717555] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 592.717660] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 592.759212] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] [instance: d75e4be8-243b-40ee-9b69-b64e0f315700] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 592.759911] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 592.760149] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 592.760323] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 592.763390] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 592.763390] nova-conductor[52523]: Traceback (most recent call last): [ 592.763390] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 592.763390] nova-conductor[52523]: return func(*args, **kwargs) [ 592.763390] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 592.763390] nova-conductor[52523]: selections = self._select_destinations( [ 592.763390] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 592.763390] nova-conductor[52523]: selections = self._schedule( [ 592.763390] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 592.763390] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 592.763390] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 592.763390] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 592.763390] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 592.763390] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 592.763910] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-a10c4d7b-77b3-4fae-994c-b46bf71c4e52 tempest-ServersWithSpecificFlavorTestJSON-201170338 tempest-ServersWithSpecificFlavorTestJSON-201170338-project-member] [instance: d75e4be8-243b-40ee-9b69-b64e0f315700] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 597.009149] nova-conductor[52523]: Traceback (most recent call last): [ 597.009149] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 597.009149] nova-conductor[52523]: return func(*args, **kwargs) [ 597.009149] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 597.009149] nova-conductor[52523]: selections = self._select_destinations( [ 597.009149] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 597.009149] nova-conductor[52523]: selections = self._schedule( [ 597.009149] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 597.009149] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 597.009149] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 597.009149] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 597.009149] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 597.009149] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 597.010074] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.010705] nova-conductor[52523]: ERROR nova.conductor.manager [ 597.017958] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 597.019242] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 597.019242] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 597.070196] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] [instance: 7fc58bf6-ad69-4607-8f3f-5e8f70bf98f4] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 597.070976] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 597.071211] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 597.071385] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 597.075027] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 597.075027] nova-conductor[52523]: Traceback (most recent call last): [ 597.075027] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 597.075027] nova-conductor[52523]: return func(*args, **kwargs) [ 597.075027] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 597.075027] nova-conductor[52523]: selections = self._select_destinations( [ 597.075027] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 597.075027] nova-conductor[52523]: selections = self._schedule( [ 597.075027] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 597.075027] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 597.075027] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 597.075027] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 597.075027] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 597.075027] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 597.075691] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-5c5575c0-3e18-4f9e-9afa-8d5eaab7a04f tempest-ServersAdminNegativeTestJSON-475667046 tempest-ServersAdminNegativeTestJSON-475667046-project-member] [instance: 7fc58bf6-ad69-4607-8f3f-5e8f70bf98f4] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 602.344305] nova-conductor[52522]: Traceback (most recent call last): [ 602.344305] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 602.344305] nova-conductor[52522]: return func(*args, **kwargs) [ 602.344305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 602.344305] nova-conductor[52522]: selections = self._select_destinations( [ 602.344305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 602.344305] nova-conductor[52522]: selections = self._schedule( [ 602.344305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 602.344305] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 602.344305] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 602.344305] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 602.344305] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 602.344305] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 602.346689] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.347515] nova-conductor[52522]: ERROR nova.conductor.manager [ 602.354208] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 602.354208] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 602.354208] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 602.403972] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] [instance: 10bdea6c-7b23-4281-ac02-4d56de8e1187] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 602.404726] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 602.404943] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 602.405146] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 602.411207] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 602.411207] nova-conductor[52522]: Traceback (most recent call last): [ 602.411207] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 602.411207] nova-conductor[52522]: return func(*args, **kwargs) [ 602.411207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 602.411207] nova-conductor[52522]: selections = self._select_destinations( [ 602.411207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 602.411207] nova-conductor[52522]: selections = self._schedule( [ 602.411207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 602.411207] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 602.411207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 602.411207] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 602.411207] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 602.411207] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 602.411207] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-cef40b5b-d846-4a15-bd5c-66d750508392 tempest-ServerRescueTestJSON-69366580 tempest-ServerRescueTestJSON-69366580-project-member] [instance: 10bdea6c-7b23-4281-ac02-4d56de8e1187] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 605.643942] nova-conductor[52523]: Traceback (most recent call last): [ 605.643942] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 605.643942] nova-conductor[52523]: return func(*args, **kwargs) [ 605.643942] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 605.643942] nova-conductor[52523]: selections = self._select_destinations( [ 605.643942] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 605.643942] nova-conductor[52523]: selections = self._schedule( [ 605.643942] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 605.643942] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 605.643942] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 605.643942] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 605.643942] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 605.643942] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 605.644934] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.648572] nova-conductor[52523]: ERROR nova.conductor.manager [ 605.655029] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 605.655324] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 605.655515] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 605.708458] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] [instance: afe5dac8-5ba5-4845-847b-4357f436c4c7] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 605.709201] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 605.709417] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 605.709585] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 605.712568] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 605.712568] nova-conductor[52523]: Traceback (most recent call last): [ 605.712568] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 605.712568] nova-conductor[52523]: return func(*args, **kwargs) [ 605.712568] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 605.712568] nova-conductor[52523]: selections = self._select_destinations( [ 605.712568] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 605.712568] nova-conductor[52523]: selections = self._schedule( [ 605.712568] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 605.712568] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 605.712568] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 605.712568] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 605.712568] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 605.712568] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 605.713151] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-6bb08bc4-0a0c-41e6-ad23-c2abfa33542e tempest-MigrationsAdminTest-1151403404 tempest-MigrationsAdminTest-1151403404-project-member] [instance: afe5dac8-5ba5-4845-847b-4357f436c4c7] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 608.628182] nova-conductor[52523]: ERROR nova.scheduler.utils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] [instance: 15e44d1f-ae9b-4ff7-841c-90acc81cf38b] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 15e44d1f-ae9b-4ff7-841c-90acc81cf38b was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 608.633574] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Rescheduling: True {{(pid=52523) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 608.633979] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 15e44d1f-ae9b-4ff7-841c-90acc81cf38b.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 15e44d1f-ae9b-4ff7-841c-90acc81cf38b. [ 608.634226] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-6dd9723a-6ac3-4c93-9836-da2fae0413c3 tempest-ServerDiagnosticsV248Test-927759409 tempest-ServerDiagnosticsV248Test-927759409-project-member] [instance: 15e44d1f-ae9b-4ff7-841c-90acc81cf38b] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 15e44d1f-ae9b-4ff7-841c-90acc81cf38b. [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 611.655075] nova-conductor[52523]: Traceback (most recent call last): [ 611.655075] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 611.655075] nova-conductor[52523]: return func(*args, **kwargs) [ 611.655075] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 611.655075] nova-conductor[52523]: selections = self._select_destinations( [ 611.655075] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 611.655075] nova-conductor[52523]: selections = self._schedule( [ 611.655075] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 611.655075] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 611.655075] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 611.655075] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 611.655075] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.655075] nova-conductor[52523]: ERROR nova.conductor.manager [ 611.662755] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 611.663437] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 611.663437] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 611.734495] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] [instance: 8d6a65b2-3437-470d-9bfa-e3fd914c9da4] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 611.734495] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 611.734643] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 611.734814] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 611.739165] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 611.739165] nova-conductor[52523]: Traceback (most recent call last): [ 611.739165] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 611.739165] nova-conductor[52523]: return func(*args, **kwargs) [ 611.739165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 611.739165] nova-conductor[52523]: selections = self._select_destinations( [ 611.739165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 611.739165] nova-conductor[52523]: selections = self._schedule( [ 611.739165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 611.739165] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 611.739165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 611.739165] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 611.739165] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 611.739165] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 611.739718] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-dbdc047b-27e4-43fb-9ec5-be626294edb3 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] [instance: 8d6a65b2-3437-470d-9bfa-e3fd914c9da4] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 615.128951] nova-conductor[52522]: Traceback (most recent call last): [ 615.128951] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 615.128951] nova-conductor[52522]: return func(*args, **kwargs) [ 615.128951] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 615.128951] nova-conductor[52522]: selections = self._select_destinations( [ 615.128951] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 615.128951] nova-conductor[52522]: selections = self._schedule( [ 615.128951] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 615.128951] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 615.128951] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 615.128951] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 615.128951] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.128951] nova-conductor[52522]: ERROR nova.conductor.manager [ 615.136799] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 615.138580] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 615.138580] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 615.213195] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] [instance: 899e626a-cc74-4991-9dd5-aab0782e7140] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 615.214166] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 615.215259] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 615.215259] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 615.218317] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 615.218317] nova-conductor[52522]: Traceback (most recent call last): [ 615.218317] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 615.218317] nova-conductor[52522]: return func(*args, **kwargs) [ 615.218317] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 615.218317] nova-conductor[52522]: selections = self._select_destinations( [ 615.218317] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 615.218317] nova-conductor[52522]: selections = self._schedule( [ 615.218317] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 615.218317] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 615.218317] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 615.218317] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 615.218317] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 615.218317] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 615.219898] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-34c92241-58f3-491e-9c93-cbc43e2f58c7 tempest-ServerRescueTestJSONUnderV235-142262656 tempest-ServerRescueTestJSONUnderV235-142262656-project-member] [instance: 899e626a-cc74-4991-9dd5-aab0782e7140] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 616.800974] nova-conductor[52522]: Traceback (most recent call last): [ 616.800974] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 616.800974] nova-conductor[52522]: return func(*args, **kwargs) [ 616.800974] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 616.800974] nova-conductor[52522]: selections = self._select_destinations( [ 616.800974] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 616.800974] nova-conductor[52522]: selections = self._schedule( [ 616.800974] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 616.800974] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 616.800974] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 616.800974] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 616.800974] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.800974] nova-conductor[52522]: ERROR nova.conductor.manager [ 616.818346] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 616.818346] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 616.818346] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 616.913021] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] [instance: ca41f71e-eb90-4e0a-ac42-2122e810f7d8] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 616.913021] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 616.913021] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 616.913021] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 616.919909] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 616.919909] nova-conductor[52522]: Traceback (most recent call last): [ 616.919909] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 616.919909] nova-conductor[52522]: return func(*args, **kwargs) [ 616.919909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 616.919909] nova-conductor[52522]: selections = self._select_destinations( [ 616.919909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 616.919909] nova-conductor[52522]: selections = self._schedule( [ 616.919909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 616.919909] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 616.919909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 616.919909] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 616.919909] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 616.919909] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 616.920758] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-8421f355-708b-489a-8b22-cf019fd78c3c tempest-ServerMetadataTestJSON-8839926 tempest-ServerMetadataTestJSON-8839926-project-member] [instance: ca41f71e-eb90-4e0a-ac42-2122e810f7d8] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.145331] nova-conductor[52523]: Traceback (most recent call last): [ 617.145331] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 617.145331] nova-conductor[52523]: return func(*args, **kwargs) [ 617.145331] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 617.145331] nova-conductor[52523]: selections = self._select_destinations( [ 617.145331] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 617.145331] nova-conductor[52523]: selections = self._schedule( [ 617.145331] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 617.145331] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 617.145331] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 617.145331] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 617.145331] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.145331] nova-conductor[52523]: ERROR nova.conductor.manager [ 617.154312] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 617.154312] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 617.154312] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 617.291737] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] [instance: cc61c6f7-af2e-4bab-bc40-45630b300593] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 617.296927] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 617.296927] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 617.296927] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 617.302846] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 617.302846] nova-conductor[52523]: Traceback (most recent call last): [ 617.302846] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 617.302846] nova-conductor[52523]: return func(*args, **kwargs) [ 617.302846] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 617.302846] nova-conductor[52523]: selections = self._select_destinations( [ 617.302846] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 617.302846] nova-conductor[52523]: selections = self._schedule( [ 617.302846] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 617.302846] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 617.302846] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 617.302846] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 617.302846] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 617.302846] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.303520] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-4b229b9a-f61f-4ae7-a7f9-0daf879c82b1 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] [instance: cc61c6f7-af2e-4bab-bc40-45630b300593] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.747909] nova-conductor[52522]: Traceback (most recent call last): [ 617.747909] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 617.747909] nova-conductor[52522]: return func(*args, **kwargs) [ 617.747909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 617.747909] nova-conductor[52522]: selections = self._select_destinations( [ 617.747909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 617.747909] nova-conductor[52522]: selections = self._schedule( [ 617.747909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 617.747909] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 617.747909] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 617.747909] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 617.747909] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.747909] nova-conductor[52522]: ERROR nova.conductor.manager [ 617.755613] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 617.755764] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 617.755944] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 617.819508] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] [instance: a6f191c5-287b-4cfa-91a6-4e39150af69f] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 617.820273] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 617.820619] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 617.820987] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 617.825290] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 617.825290] nova-conductor[52522]: Traceback (most recent call last): [ 617.825290] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 617.825290] nova-conductor[52522]: return func(*args, **kwargs) [ 617.825290] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 617.825290] nova-conductor[52522]: selections = self._select_destinations( [ 617.825290] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 617.825290] nova-conductor[52522]: selections = self._schedule( [ 617.825290] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 617.825290] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 617.825290] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 617.825290] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 617.825290] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 617.825290] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 617.825904] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-89845ad9-b946-429b-b120-30d97a78d644 tempest-ServerDiagnosticsTest-1335090358 tempest-ServerDiagnosticsTest-1335090358-project-member] [instance: a6f191c5-287b-4cfa-91a6-4e39150af69f] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 619.827269] nova-conductor[52523]: Traceback (most recent call last): [ 619.827269] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 619.827269] nova-conductor[52523]: return func(*args, **kwargs) [ 619.827269] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 619.827269] nova-conductor[52523]: selections = self._select_destinations( [ 619.827269] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 619.827269] nova-conductor[52523]: selections = self._schedule( [ 619.827269] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 619.827269] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 619.827269] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 619.827269] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 619.827269] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.827269] nova-conductor[52523]: ERROR nova.conductor.manager [ 619.838990] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 619.839406] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 619.839665] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 619.921267] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] [instance: 7ccbc341-996e-4730-a74d-43c7155af4d4] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 619.922115] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 619.922414] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 619.922621] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 619.926035] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 619.926035] nova-conductor[52523]: Traceback (most recent call last): [ 619.926035] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 619.926035] nova-conductor[52523]: return func(*args, **kwargs) [ 619.926035] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 619.926035] nova-conductor[52523]: selections = self._select_destinations( [ 619.926035] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 619.926035] nova-conductor[52523]: selections = self._schedule( [ 619.926035] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 619.926035] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 619.926035] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 619.926035] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 619.926035] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 619.926035] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 619.927189] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-1fec61a1-bb77-44d0-8034-9a760afa50ed tempest-ServersV294TestFqdnHostnames-559436711 tempest-ServersV294TestFqdnHostnames-559436711-project-member] [instance: 7ccbc341-996e-4730-a74d-43c7155af4d4] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.146134] nova-conductor[52522]: Traceback (most recent call last): [ 621.146134] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 621.146134] nova-conductor[52522]: return func(*args, **kwargs) [ 621.146134] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 621.146134] nova-conductor[52522]: selections = self._select_destinations( [ 621.146134] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 621.146134] nova-conductor[52522]: selections = self._schedule( [ 621.146134] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 621.146134] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 621.146134] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 621.146134] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 621.146134] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.146134] nova-conductor[52522]: ERROR nova.conductor.manager [ 621.156768] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 621.157415] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 621.157704] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 621.208462] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] [instance: 2a9a0355-c485-4613-890b-d052dcedbd28] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 621.209377] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 621.209646] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 621.209859] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 621.213311] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 621.213311] nova-conductor[52522]: Traceback (most recent call last): [ 621.213311] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 621.213311] nova-conductor[52522]: return func(*args, **kwargs) [ 621.213311] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 621.213311] nova-conductor[52522]: selections = self._select_destinations( [ 621.213311] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 621.213311] nova-conductor[52522]: selections = self._schedule( [ 621.213311] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 621.213311] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 621.213311] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 621.213311] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 621.213311] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 621.213311] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.213857] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-fe209de1-a3c4-41ed-a682-97f8ce76a2f9 tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] [instance: 2a9a0355-c485-4613-890b-d052dcedbd28] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.536721] nova-conductor[52523]: Traceback (most recent call last): [ 621.536721] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 621.536721] nova-conductor[52523]: return func(*args, **kwargs) [ 621.536721] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 621.536721] nova-conductor[52523]: selections = self._select_destinations( [ 621.536721] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 621.536721] nova-conductor[52523]: selections = self._schedule( [ 621.536721] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 621.536721] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 621.536721] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 621.536721] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 621.536721] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.536721] nova-conductor[52523]: ERROR nova.conductor.manager [ 621.544608] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 621.545370] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 621.545370] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 621.598237] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] [instance: 7f48e193-6c5e-4f5a-bf72-86dfa8a0c575] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 621.598571] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 621.598805] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 621.599076] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 621.602745] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 621.602745] nova-conductor[52523]: Traceback (most recent call last): [ 621.602745] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 621.602745] nova-conductor[52523]: return func(*args, **kwargs) [ 621.602745] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 621.602745] nova-conductor[52523]: selections = self._select_destinations( [ 621.602745] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 621.602745] nova-conductor[52523]: selections = self._schedule( [ 621.602745] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 621.602745] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 621.602745] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 621.602745] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 621.602745] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 621.602745] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 621.603285] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-d816a50b-511c-46c2-a935-95c5cdf934a4 tempest-ListServerFiltersTestJSON-1408482096 tempest-ListServerFiltersTestJSON-1408482096-project-member] [instance: 7f48e193-6c5e-4f5a-bf72-86dfa8a0c575] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 622.204031] nova-conductor[52522]: Traceback (most recent call last): [ 622.204031] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 622.204031] nova-conductor[52522]: return func(*args, **kwargs) [ 622.204031] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 622.204031] nova-conductor[52522]: selections = self._select_destinations( [ 622.204031] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 622.204031] nova-conductor[52522]: selections = self._schedule( [ 622.204031] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 622.204031] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 622.204031] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 622.204031] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 622.204031] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.204031] nova-conductor[52522]: ERROR nova.conductor.manager [ 622.213939] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 622.214009] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 622.214232] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 622.259618] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] [instance: 33a5a930-e1cb-48aa-8bcb-c5abc19ceb4c] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 622.260351] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 622.260568] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 622.261106] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 622.264826] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 622.264826] nova-conductor[52522]: Traceback (most recent call last): [ 622.264826] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 622.264826] nova-conductor[52522]: return func(*args, **kwargs) [ 622.264826] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 622.264826] nova-conductor[52522]: selections = self._select_destinations( [ 622.264826] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 622.264826] nova-conductor[52522]: selections = self._schedule( [ 622.264826] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 622.264826] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 622.264826] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 622.264826] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 622.264826] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 622.264826] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 622.265403] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-48e778ff-9564-4490-a9cf-a6e3cbbc25d4 tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] [instance: 33a5a930-e1cb-48aa-8bcb-c5abc19ceb4c] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.630731] nova-conductor[52523]: Traceback (most recent call last): [ 624.630731] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 624.630731] nova-conductor[52523]: return func(*args, **kwargs) [ 624.630731] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 624.630731] nova-conductor[52523]: selections = self._select_destinations( [ 624.630731] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 624.630731] nova-conductor[52523]: selections = self._schedule( [ 624.630731] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 624.630731] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 624.630731] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 624.630731] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 624.630731] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.630731] nova-conductor[52523]: ERROR nova.conductor.manager [ 624.638888] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 624.639725] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 624.639725] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 624.688815] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] [instance: 20811f59-6dcc-4034-a503-215eb5418062] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 624.689876] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 624.689876] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 624.690067] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 624.693479] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 624.693479] nova-conductor[52523]: Traceback (most recent call last): [ 624.693479] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 624.693479] nova-conductor[52523]: return func(*args, **kwargs) [ 624.693479] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 624.693479] nova-conductor[52523]: selections = self._select_destinations( [ 624.693479] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 624.693479] nova-conductor[52523]: selections = self._schedule( [ 624.693479] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 624.693479] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 624.693479] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 624.693479] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 624.693479] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 624.693479] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.694019] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-32e9430f-b1d0-4ab3-b0d0-23a1a8b1555d tempest-ServerRescueNegativeTestJSON-1759377260 tempest-ServerRescueNegativeTestJSON-1759377260-project-member] [instance: 20811f59-6dcc-4034-a503-215eb5418062] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.814645] nova-conductor[52522]: Traceback (most recent call last): [ 624.814645] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 624.814645] nova-conductor[52522]: return func(*args, **kwargs) [ 624.814645] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 624.814645] nova-conductor[52522]: selections = self._select_destinations( [ 624.814645] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 624.814645] nova-conductor[52522]: selections = self._schedule( [ 624.814645] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 624.814645] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 624.814645] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 624.814645] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 624.814645] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.814645] nova-conductor[52522]: ERROR nova.conductor.manager [ 624.823428] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 624.823633] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 624.823935] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 624.870431] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] [instance: 35452952-d10d-4f6b-92fa-97bb42c5bbcc] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 624.871229] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 624.871459] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 624.871632] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 624.875334] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 624.875334] nova-conductor[52522]: Traceback (most recent call last): [ 624.875334] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 624.875334] nova-conductor[52522]: return func(*args, **kwargs) [ 624.875334] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 624.875334] nova-conductor[52522]: selections = self._select_destinations( [ 624.875334] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 624.875334] nova-conductor[52522]: selections = self._schedule( [ 624.875334] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 624.875334] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 624.875334] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 624.875334] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 624.875334] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 624.875334] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 624.875927] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-a1e39e1f-fc81-4842-b673-da2d88ad1912 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] [instance: 35452952-d10d-4f6b-92fa-97bb42c5bbcc] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 627.455870] nova-conductor[52523]: Traceback (most recent call last): [ 627.455870] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 627.455870] nova-conductor[52523]: return func(*args, **kwargs) [ 627.455870] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 627.455870] nova-conductor[52523]: selections = self._select_destinations( [ 627.455870] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 627.455870] nova-conductor[52523]: selections = self._schedule( [ 627.455870] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 627.455870] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 627.455870] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 627.455870] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 627.455870] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.455870] nova-conductor[52523]: ERROR nova.conductor.manager [ 627.465445] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 627.465686] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 627.465861] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 627.535798] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] [instance: 1cf1d6fc-8bae-4797-8d7e-6ae7d62bd16f] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 627.536617] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 627.536875] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 627.537100] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 627.542221] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 627.542221] nova-conductor[52523]: Traceback (most recent call last): [ 627.542221] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 627.542221] nova-conductor[52523]: return func(*args, **kwargs) [ 627.542221] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 627.542221] nova-conductor[52523]: selections = self._select_destinations( [ 627.542221] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 627.542221] nova-conductor[52523]: selections = self._schedule( [ 627.542221] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 627.542221] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 627.542221] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 627.542221] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 627.542221] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 627.542221] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 627.542787] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-33a50cb7-baec-47eb-8eec-0fcf574636f6 tempest-ListImageFiltersTestJSON-585052884 tempest-ListImageFiltersTestJSON-585052884-project-member] [instance: 1cf1d6fc-8bae-4797-8d7e-6ae7d62bd16f] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 628.161765] nova-conductor[52522]: Traceback (most recent call last): [ 628.161765] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 628.161765] nova-conductor[52522]: return func(*args, **kwargs) [ 628.161765] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 628.161765] nova-conductor[52522]: selections = self._select_destinations( [ 628.161765] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 628.161765] nova-conductor[52522]: selections = self._schedule( [ 628.161765] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 628.161765] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 628.161765] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 628.161765] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 628.161765] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.161765] nova-conductor[52522]: ERROR nova.conductor.manager [ 628.173382] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 628.173382] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 628.173563] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 628.229212] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] [instance: b663c3c4-acdd-4153-ab0f-1cdbf6f893e6] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 628.230271] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 628.230503] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 628.230703] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 628.236071] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 628.236071] nova-conductor[52522]: Traceback (most recent call last): [ 628.236071] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 628.236071] nova-conductor[52522]: return func(*args, **kwargs) [ 628.236071] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 628.236071] nova-conductor[52522]: selections = self._select_destinations( [ 628.236071] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 628.236071] nova-conductor[52522]: selections = self._schedule( [ 628.236071] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 628.236071] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 628.236071] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 628.236071] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 628.236071] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 628.236071] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 628.237088] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-42d665c2-575d-4e1d-ba09-ef108d2a65d7 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] [instance: b663c3c4-acdd-4153-ab0f-1cdbf6f893e6] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 631.761544] nova-conductor[52523]: Traceback (most recent call last): [ 631.761544] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 631.761544] nova-conductor[52523]: return func(*args, **kwargs) [ 631.761544] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 631.761544] nova-conductor[52523]: selections = self._select_destinations( [ 631.761544] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 631.761544] nova-conductor[52523]: selections = self._schedule( [ 631.761544] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 631.761544] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 631.761544] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 631.761544] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 631.761544] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.761544] nova-conductor[52523]: ERROR nova.conductor.manager [ 631.773295] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 631.773528] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 631.773703] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 631.826399] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] [instance: 9b48df25-2dc2-423c-9408-f31388e40cd5] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 631.827113] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 631.827338] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 631.827514] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 631.830854] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 631.830854] nova-conductor[52523]: Traceback (most recent call last): [ 631.830854] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 631.830854] nova-conductor[52523]: return func(*args, **kwargs) [ 631.830854] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 631.830854] nova-conductor[52523]: selections = self._select_destinations( [ 631.830854] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 631.830854] nova-conductor[52523]: selections = self._schedule( [ 631.830854] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 631.830854] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 631.830854] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 631.830854] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 631.830854] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 631.830854] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 631.833556] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-b2f9313d-0fd0-438d-b92e-8d499dfb7d2a tempest-ServerShowV257Test-1796706101 tempest-ServerShowV257Test-1796706101-project-member] [instance: 9b48df25-2dc2-423c-9408-f31388e40cd5] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 636.352337] nova-conductor[52522]: Traceback (most recent call last): [ 636.352337] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 636.352337] nova-conductor[52522]: return func(*args, **kwargs) [ 636.352337] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 636.352337] nova-conductor[52522]: selections = self._select_destinations( [ 636.352337] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 636.352337] nova-conductor[52522]: selections = self._schedule( [ 636.352337] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 636.352337] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 636.352337] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 636.352337] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 636.352337] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.352337] nova-conductor[52522]: ERROR nova.conductor.manager [ 636.360364] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 636.360607] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 636.360808] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 636.413107] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: ade2fbd8-3ded-4c99-a189-fc885c980959] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 636.413343] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 636.413562] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 636.413736] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 636.417010] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 636.417010] nova-conductor[52522]: Traceback (most recent call last): [ 636.417010] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 636.417010] nova-conductor[52522]: return func(*args, **kwargs) [ 636.417010] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 636.417010] nova-conductor[52522]: selections = self._select_destinations( [ 636.417010] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 636.417010] nova-conductor[52522]: selections = self._schedule( [ 636.417010] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 636.417010] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 636.417010] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 636.417010] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 636.417010] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 636.417010] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 636.417543] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: ade2fbd8-3ded-4c99-a189-fc885c980959] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 636.472147] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 636.472396] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 636.472573] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 636.541830] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: 7b4135e7-c0e2-4ab7-8cba-19d2059e40e2] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 636.542589] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 636.542806] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 636.542979] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 636.552472] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 636.552472] nova-conductor[52522]: Traceback (most recent call last): [ 636.552472] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 636.552472] nova-conductor[52522]: return func(*args, **kwargs) [ 636.552472] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 636.552472] nova-conductor[52522]: selections = self._select_destinations( [ 636.552472] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 636.552472] nova-conductor[52522]: selections = self._schedule( [ 636.552472] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 636.552472] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 636.552472] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 636.552472] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 636.552472] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 636.552472] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 636.552472] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-823b1ff9-6c49-44df-a391-d35bcd60e92e tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: 7b4135e7-c0e2-4ab7-8cba-19d2059e40e2] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 639.971112] nova-conductor[52523]: Traceback (most recent call last): [ 639.971112] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 639.971112] nova-conductor[52523]: return func(*args, **kwargs) [ 639.971112] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 639.971112] nova-conductor[52523]: selections = self._select_destinations( [ 639.971112] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 639.971112] nova-conductor[52523]: selections = self._schedule( [ 639.971112] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 639.971112] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 639.971112] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 639.971112] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 639.971112] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.971112] nova-conductor[52523]: ERROR nova.conductor.manager [ 639.981186] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 639.981447] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 639.981626] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 640.044254] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] [instance: 0a342770-1daf-41d6-9796-a927458ee8ca] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 640.046654] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 640.046892] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 640.047077] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 640.050405] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 640.050405] nova-conductor[52523]: Traceback (most recent call last): [ 640.050405] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 640.050405] nova-conductor[52523]: return func(*args, **kwargs) [ 640.050405] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 640.050405] nova-conductor[52523]: selections = self._select_destinations( [ 640.050405] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 640.050405] nova-conductor[52523]: selections = self._schedule( [ 640.050405] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 640.050405] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 640.050405] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 640.050405] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 640.050405] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 640.050405] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 640.051489] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-ab07824d-c168-4d37-94d9-691f9fb763df tempest-AttachVolumeShelveTestJSON-1407032086 tempest-AttachVolumeShelveTestJSON-1407032086-project-member] [instance: 0a342770-1daf-41d6-9796-a927458ee8ca] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 640.401151] nova-conductor[52522]: Traceback (most recent call last): [ 640.401151] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 640.401151] nova-conductor[52522]: return func(*args, **kwargs) [ 640.401151] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 640.401151] nova-conductor[52522]: selections = self._select_destinations( [ 640.401151] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 640.401151] nova-conductor[52522]: selections = self._schedule( [ 640.401151] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 640.401151] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 640.401151] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 640.401151] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 640.401151] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.401151] nova-conductor[52522]: ERROR nova.conductor.manager [ 640.410021] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 640.410021] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 640.410208] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 640.462635] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] [instance: f5ab17f1-ce4d-4437-b2bb-8860db593937] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 640.463386] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 640.463608] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 640.463781] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 640.467207] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 640.467207] nova-conductor[52522]: Traceback (most recent call last): [ 640.467207] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 640.467207] nova-conductor[52522]: return func(*args, **kwargs) [ 640.467207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 640.467207] nova-conductor[52522]: selections = self._select_destinations( [ 640.467207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 640.467207] nova-conductor[52522]: selections = self._schedule( [ 640.467207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 640.467207] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 640.467207] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 640.467207] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 640.467207] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 640.467207] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 640.468085] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-eb765f79-d3c7-4a0a-bbd3-2d60a3bc4a06 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] [instance: f5ab17f1-ce4d-4437-b2bb-8860db593937] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 641.602668] nova-conductor[52523]: Traceback (most recent call last): [ 641.602668] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 641.602668] nova-conductor[52523]: return func(*args, **kwargs) [ 641.602668] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 641.602668] nova-conductor[52523]: selections = self._select_destinations( [ 641.602668] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 641.602668] nova-conductor[52523]: selections = self._schedule( [ 641.602668] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 641.602668] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 641.602668] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 641.602668] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 641.602668] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.602668] nova-conductor[52523]: ERROR nova.conductor.manager [ 641.619068] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 641.619068] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 641.619068] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 641.689394] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] [instance: e51c23a2-cba1-4847-bb60-4adae4d3a8d9] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 641.689394] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 641.689614] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 641.689614] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 641.693136] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 641.693136] nova-conductor[52523]: Traceback (most recent call last): [ 641.693136] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 641.693136] nova-conductor[52523]: return func(*args, **kwargs) [ 641.693136] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 641.693136] nova-conductor[52523]: selections = self._select_destinations( [ 641.693136] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 641.693136] nova-conductor[52523]: selections = self._schedule( [ 641.693136] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 641.693136] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 641.693136] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 641.693136] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 641.693136] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 641.693136] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 641.693644] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-59082ba2-0f34-4f2b-9617-0e114f2500e1 tempest-VolumesAdminNegativeTest-1936733797 tempest-VolumesAdminNegativeTest-1936733797-project-member] [instance: e51c23a2-cba1-4847-bb60-4adae4d3a8d9] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 643.961016] nova-conductor[52522]: Traceback (most recent call last): [ 643.961016] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 643.961016] nova-conductor[52522]: return func(*args, **kwargs) [ 643.961016] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 643.961016] nova-conductor[52522]: selections = self._select_destinations( [ 643.961016] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 643.961016] nova-conductor[52522]: selections = self._schedule( [ 643.961016] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 643.961016] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 643.961016] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 643.961016] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 643.961016] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.961016] nova-conductor[52522]: ERROR nova.conductor.manager [ 643.970864] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 643.972101] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 643.972101] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 644.012663] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] [instance: 6cb9dbfc-3cfe-4769-9deb-506c3a540b2d] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 644.013439] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 644.013731] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 644.014241] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 644.020584] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 644.020584] nova-conductor[52522]: Traceback (most recent call last): [ 644.020584] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 644.020584] nova-conductor[52522]: return func(*args, **kwargs) [ 644.020584] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 644.020584] nova-conductor[52522]: selections = self._select_destinations( [ 644.020584] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 644.020584] nova-conductor[52522]: selections = self._schedule( [ 644.020584] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 644.020584] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 644.020584] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 644.020584] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 644.020584] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 644.020584] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 644.021427] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-72ddf0c8-7ebf-42d2-928b-7f2d0e9a44d2 tempest-ServersAaction247Test-191636729 tempest-ServersAaction247Test-191636729-project-member] [instance: 6cb9dbfc-3cfe-4769-9deb-506c3a540b2d] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 644.584860] nova-conductor[52523]: Traceback (most recent call last): [ 644.584860] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 644.584860] nova-conductor[52523]: return func(*args, **kwargs) [ 644.584860] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 644.584860] nova-conductor[52523]: selections = self._select_destinations( [ 644.584860] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 644.584860] nova-conductor[52523]: selections = self._schedule( [ 644.584860] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 644.584860] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 644.584860] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 644.584860] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 644.584860] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.584860] nova-conductor[52523]: ERROR nova.conductor.manager [ 644.594217] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 644.594217] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 644.594217] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 644.647347] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] [instance: 4e8163e4-0872-44f1-bb13-fab896e4dc4d] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 644.648218] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 644.648854] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 644.649146] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 644.652722] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 644.652722] nova-conductor[52523]: Traceback (most recent call last): [ 644.652722] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 644.652722] nova-conductor[52523]: return func(*args, **kwargs) [ 644.652722] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 644.652722] nova-conductor[52523]: selections = self._select_destinations( [ 644.652722] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 644.652722] nova-conductor[52523]: selections = self._schedule( [ 644.652722] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 644.652722] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 644.652722] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 644.652722] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 644.652722] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 644.652722] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 644.653920] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-59138858-f9b9-40b5-b52d-4b24065b7231 tempest-ServerTagsTestJSON-1761503152 tempest-ServerTagsTestJSON-1761503152-project-member] [instance: 4e8163e4-0872-44f1-bb13-fab896e4dc4d] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 651.333631] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Took 0.19 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 651.352285] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 651.352471] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 651.353189] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 651.407856] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 651.408154] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 651.408207] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 651.408629] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 651.408776] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 651.408889] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 651.423678] nova-conductor[52523]: DEBUG nova.quota [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Getting quotas for project f42f07b8ce5040279223112427eb62c9. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 651.426265] nova-conductor[52523]: DEBUG nova.quota [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Getting quotas for user 41a535ff3d8148c29a9bf62afbeff688 and project f42f07b8ce5040279223112427eb62c9. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 651.438202] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] [instance: 08e2d758-9005-4822-b157-84710b9c5ed4] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 651.438573] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 651.438827] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 651.438963] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 651.443985] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] [instance: 08e2d758-9005-4822-b157-84710b9c5ed4] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 651.444854] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 651.445150] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 651.445374] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 651.471026] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 651.471099] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 651.471255] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 652.888238] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Took 0.23 seconds to select destinations for 2 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 652.901238] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 652.901502] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 652.901681] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 652.931569] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 652.931806] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 652.931982] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 652.956225] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 652.956428] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 652.956682] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 652.956969] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 652.957167] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 652.957343] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 652.966041] nova-conductor[52522]: DEBUG nova.quota [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Getting quotas for project 6c6cba030fa2464f98c773682138ae9c. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 652.969174] nova-conductor[52522]: DEBUG nova.quota [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Getting quotas for user ff2598db3b974d7685d57094808f2ef8 and project 6c6cba030fa2464f98c773682138ae9c. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 652.974498] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: 54919bf0-b9f3-4bfc-ba1a-c6a52013e351] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 652.975151] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 652.975392] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 652.975581] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 652.984955] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: 54919bf0-b9f3-4bfc-ba1a-c6a52013e351] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 652.985728] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 652.985931] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 652.986118] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.001900] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.002145] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.002343] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.008627] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: a925d5fc-6437-40bb-adf1-ea10c32dde2a] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 653.009274] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.009375] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.009547] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.013424] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] [instance: a925d5fc-6437-40bb-adf1-ea10c32dde2a] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 653.013550] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.013779] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.013950] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.028841] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.029140] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.029479] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.863723] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Took 0.17 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 653.880476] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.880900] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.883023] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.940033] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.940033] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.940233] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.940491] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.941346] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.941346] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.952360] nova-conductor[52523]: DEBUG nova.quota [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Getting quotas for project af470ef5e36145e3bb547ba685209d97. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 653.955304] nova-conductor[52523]: DEBUG nova.quota [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Getting quotas for user 1cdb31acf24540e18e093a55808b2a84 and project af470ef5e36145e3bb547ba685209d97. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 653.962380] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] [instance: 076c3dd5-9043-456d-af24-0d2273321085] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 653.962380] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.962380] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.962612] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.966795] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] [instance: 076c3dd5-9043-456d-af24-0d2273321085] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 653.968026] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.968026] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.968026] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 653.987289] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 653.988605] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 653.988605] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 658.469184] nova-conductor[52523]: ERROR nova.scheduler.utils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 8a0d9829-6759-4593-9230-459a546a5908] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 8a0d9829-6759-4593-9230-459a546a5908 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 658.470221] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Rescheduling: True {{(pid=52523) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 658.470493] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 8a0d9829-6759-4593-9230-459a546a5908.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 8a0d9829-6759-4593-9230-459a546a5908. [ 658.470830] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 8a0d9829-6759-4593-9230-459a546a5908] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 8a0d9829-6759-4593-9230-459a546a5908. [ 658.506354] nova-conductor[52523]: DEBUG nova.network.neutron [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 8a0d9829-6759-4593-9230-459a546a5908] deallocate_for_instance() {{(pid=52523) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.905091] nova-conductor[52523]: DEBUG nova.network.neutron [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 8a0d9829-6759-4593-9230-459a546a5908] Instance cache missing network info. {{(pid=52523) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.916104] nova-conductor[52523]: DEBUG nova.network.neutron [None req-b44b385c-ccd6-4d8e-a3a7-e113eb34c2f3 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 8a0d9829-6759-4593-9230-459a546a5908] Updating instance_info_cache with network_info: [] {{(pid=52523) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.415549] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Took 0.16 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 661.438479] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 661.438747] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 661.438926] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 661.471377] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 661.471628] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 661.471808] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 661.472191] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 661.472406] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 661.472559] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 661.482603] nova-conductor[52523]: DEBUG nova.quota [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Getting quotas for project 8d5c11c9c30b43b9964cbba3f0e5ebb3. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 661.488028] nova-conductor[52523]: DEBUG nova.quota [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Getting quotas for user 680d8f0fb2af458fb4ffb6f070b89813 and project 8d5c11c9c30b43b9964cbba3f0e5ebb3. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 661.493249] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] [instance: 1eaf8e02-bfb0-4928-9687-cc781a84d16d] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 661.493827] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 661.494071] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 661.495045] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 661.501511] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] [instance: 1eaf8e02-bfb0-4928-9687-cc781a84d16d] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 661.502304] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 661.502513] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 661.502699] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 661.527365] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 661.527617] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 661.528246] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 663.796833] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Took 0.16 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 663.812082] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 663.812487] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.004s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 663.812722] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 663.847146] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 663.847383] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 663.847566] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 663.847924] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 663.848125] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 663.848290] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 663.861203] nova-conductor[52523]: DEBUG nova.quota [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Getting quotas for project 4b3edaaa3fdc4f73b49b8e57e04b8fa0. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 663.863309] nova-conductor[52523]: DEBUG nova.quota [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Getting quotas for user 5208b8d162bc46d489a34997aaebbaa2 and project 4b3edaaa3fdc4f73b49b8e57e04b8fa0. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 663.874196] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: c841298b-f103-4dc7-8884-efdf2ebc20a6] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 663.874196] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 663.874196] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 663.874196] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 663.876582] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: c841298b-f103-4dc7-8884-efdf2ebc20a6] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 663.877512] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 663.877720] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 663.877883] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 663.892409] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 663.893354] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 663.893587] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 666.413309] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Took 0.14 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 666.428226] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 666.428475] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 666.428651] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 666.460565] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 666.460805] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 666.460974] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 666.461497] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 666.461694] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 666.461861] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 666.472927] nova-conductor[52522]: DEBUG nova.quota [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Getting quotas for project 1a960230130c47b8b17bef2a06fe0089. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 666.475945] nova-conductor[52522]: DEBUG nova.quota [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Getting quotas for user bbeb5e48a76f42449097ca4fc9a18824 and project 1a960230130c47b8b17bef2a06fe0089. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 666.482566] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: bf70d23b-4ab5-476e-814c-264b6a9f2455] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 666.483086] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 666.483278] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 666.483451] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 666.486610] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: bf70d23b-4ab5-476e-814c-264b6a9f2455] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 666.487288] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 666.487494] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 666.487667] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 666.499828] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 666.500053] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 666.500230] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.323786] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Took 0.13 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 669.335758] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.336008] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.336205] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.362577] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.362826] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.363012] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.363380] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.363569] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.363733] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.372365] nova-conductor[52523]: DEBUG nova.quota [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Getting quotas for project 8b63998428d64a5eb592ed2deaaaf649. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 669.375283] nova-conductor[52523]: DEBUG nova.quota [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Getting quotas for user 626ea401c74540919836a15d62ad4d49 and project 8b63998428d64a5eb592ed2deaaaf649. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 669.381370] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] [instance: 67ff7d52-6e30-4730-9b5a-9ae32f68b953] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 669.381872] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.382099] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.382279] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.385096] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] [instance: 67ff7d52-6e30-4730-9b5a-9ae32f68b953] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 669.385775] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.385979] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.386166] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.405821] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.406066] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.406980] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.540579] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Took 0.14 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 669.552703] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.552944] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.553135] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.584627] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.584781] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.584955] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.585311] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.585495] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.585659] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.594149] nova-conductor[52522]: DEBUG nova.quota [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Getting quotas for project d6c52e03a9e243d39ffc8dc15485ba86. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 669.596528] nova-conductor[52522]: DEBUG nova.quota [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Getting quotas for user bec422b47efa4ddc804e737c854f3f46 and project d6c52e03a9e243d39ffc8dc15485ba86. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 669.602666] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] [instance: bd91e947-acae-4dbd-b48b-5a6727eb4cbb] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 669.603190] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.603398] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.603571] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.606793] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] [instance: bd91e947-acae-4dbd-b48b-5a6727eb4cbb] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 669.608411] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.608411] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.608411] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 669.621407] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 669.621712] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 669.621831] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 677.338787] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Took 0.14 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 677.353756] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 677.354089] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 677.354322] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 677.388037] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 677.388370] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 677.388625] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 677.389057] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 677.389356] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 677.389590] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 677.398969] nova-conductor[52523]: DEBUG nova.quota [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Getting quotas for project af76eb44cf9f4a32a397133d730cecd3. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 677.401428] nova-conductor[52523]: DEBUG nova.quota [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Getting quotas for user 1ec4e5b66b3a4092851ef66e913dd8b2 and project af76eb44cf9f4a32a397133d730cecd3. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 677.407609] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] [instance: 95358801-c9d8-4582-a712-36a8bf586456] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 677.408292] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 677.408627] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 677.408931] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 677.412356] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] [instance: 95358801-c9d8-4582-a712-36a8bf586456] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 677.413865] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 677.414180] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 677.414400] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 677.428416] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 677.428642] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 677.428832] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 694.396550] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Took 0.12 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 694.409363] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 694.409621] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 694.409806] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 694.439645] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 694.439933] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 694.440135] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 694.440576] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 694.440774] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 694.440945] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 694.449727] nova-conductor[52522]: DEBUG nova.quota [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Getting quotas for project 4fb067122f8242dbbd0ae6937663f1bd. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 694.452304] nova-conductor[52522]: DEBUG nova.quota [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Getting quotas for user a0f3d43bbbcd49ee95648cb503cfa1f1 and project 4fb067122f8242dbbd0ae6937663f1bd. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 694.458086] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] [instance: be6a4290-dbb3-4e1f-bdd4-0dc106db9435] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 694.458590] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 694.458805] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 694.459015] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 694.461922] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] [instance: be6a4290-dbb3-4e1f-bdd4-0dc106db9435] block_device_mapping [BlockDeviceMapping(attachment_id=13f0b037-368e-4a4f-a41c-bc1a23d7d687,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='volume',device_name=None,device_type=None,disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id=None,instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='volume',tag=None,updated_at=,uuid=,volume_id='8d6691bf-275d-4616-badd-c8004e156406',volume_size=1,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 694.462638] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 694.462847] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 694.463029] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 694.477033] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 694.477033] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 694.477033] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 705.474295] nova-conductor[52523]: ERROR nova.scheduler.utils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] [instance: 68c87b51-b90a-47cc-bec1-05f7c389fc14] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 68c87b51-b90a-47cc-bec1-05f7c389fc14 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 705.474944] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Rescheduling: True {{(pid=52523) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 705.475131] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 68c87b51-b90a-47cc-bec1-05f7c389fc14.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 68c87b51-b90a-47cc-bec1-05f7c389fc14. [ 705.475477] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-b2f2d782-6e13-439a-b7a6-82d4614f8bd4 tempest-ServersAdmin275Test-1140959763 tempest-ServersAdmin275Test-1140959763-project-member] [instance: 68c87b51-b90a-47cc-bec1-05f7c389fc14] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 68c87b51-b90a-47cc-bec1-05f7c389fc14. [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 710.491126] nova-conductor[52523]: Traceback (most recent call last): [ 710.491126] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 710.491126] nova-conductor[52523]: return func(*args, **kwargs) [ 710.491126] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 710.491126] nova-conductor[52523]: selections = self._select_destinations( [ 710.491126] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 710.491126] nova-conductor[52523]: selections = self._schedule( [ 710.491126] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 710.491126] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 710.491126] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 710.491126] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 710.491126] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.491126] nova-conductor[52523]: ERROR nova.conductor.manager [ 710.509057] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 710.509057] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 710.509057] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 710.560078] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] [instance: 9f2f3b9b-01cf-4709-acc0-e56e1951dd7a] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 710.560742] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 710.560961] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 710.561155] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 710.564391] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 710.564391] nova-conductor[52523]: Traceback (most recent call last): [ 710.564391] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 710.564391] nova-conductor[52523]: return func(*args, **kwargs) [ 710.564391] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 710.564391] nova-conductor[52523]: selections = self._select_destinations( [ 710.564391] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 710.564391] nova-conductor[52523]: selections = self._schedule( [ 710.564391] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 710.564391] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 710.564391] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 710.564391] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 710.564391] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 710.564391] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 710.564953] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-1172f4db-615e-4e7d-966c-dc753f585e0b tempest-ServerMetadataNegativeTestJSON-1998823510 tempest-ServerMetadataNegativeTestJSON-1998823510-project-member] [instance: 9f2f3b9b-01cf-4709-acc0-e56e1951dd7a] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 753.386034] nova-conductor[52522]: ERROR nova.scheduler.utils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] [instance: c4400e80-4457-4a8a-8588-f594e5993cde] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance c4400e80-4457-4a8a-8588-f594e5993cde was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 753.386034] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Rescheduling: True {{(pid=52522) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 753.386034] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance c4400e80-4457-4a8a-8588-f594e5993cde.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance c4400e80-4457-4a8a-8588-f594e5993cde. [ 753.386034] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] [instance: c4400e80-4457-4a8a-8588-f594e5993cde] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance c4400e80-4457-4a8a-8588-f594e5993cde. [ 753.428630] nova-conductor[52522]: DEBUG nova.network.neutron [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] [instance: c4400e80-4457-4a8a-8588-f594e5993cde] deallocate_for_instance() {{(pid=52522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.568280] nova-conductor[52522]: DEBUG nova.network.neutron [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] [instance: c4400e80-4457-4a8a-8588-f594e5993cde] Instance cache missing network info. {{(pid=52522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.575505] nova-conductor[52522]: DEBUG nova.network.neutron [None req-268f4dc5-3f67-4434-af1a-623b52e88f81 tempest-ServerDiagnosticsNegativeTest-1478228549 tempest-ServerDiagnosticsNegativeTest-1478228549-project-member] [instance: c4400e80-4457-4a8a-8588-f594e5993cde] Updating instance_info_cache with network_info: [] {{(pid=52522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 758.920099] nova-conductor[52523]: Traceback (most recent call last): [ 758.920099] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 758.920099] nova-conductor[52523]: return func(*args, **kwargs) [ 758.920099] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 758.920099] nova-conductor[52523]: selections = self._select_destinations( [ 758.920099] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 758.920099] nova-conductor[52523]: selections = self._schedule( [ 758.920099] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 758.920099] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 758.920099] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 758.920099] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 758.920099] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.920099] nova-conductor[52523]: ERROR nova.conductor.manager [ 758.936020] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 758.936020] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 758.936020] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 758.980201] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] [instance: 79f4bcb9-69b1-4084-bdf9-92b90f09c3b6] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 758.980946] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 758.981177] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 758.981359] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 758.984785] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 758.984785] nova-conductor[52523]: Traceback (most recent call last): [ 758.984785] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 758.984785] nova-conductor[52523]: return func(*args, **kwargs) [ 758.984785] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 758.984785] nova-conductor[52523]: selections = self._select_destinations( [ 758.984785] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 758.984785] nova-conductor[52523]: selections = self._schedule( [ 758.984785] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 758.984785] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 758.984785] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 758.984785] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 758.984785] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 758.984785] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 758.985854] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-34717cf9-ea56-4e13-92fc-747e0177547c tempest-ImagesOneServerNegativeTestJSON-293876941 tempest-ImagesOneServerNegativeTestJSON-293876941-project-member] [instance: 79f4bcb9-69b1-4084-bdf9-92b90f09c3b6] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 767.419051] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Took 0.13 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 767.431700] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 767.431938] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 767.432177] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 767.467651] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 767.467887] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 767.468074] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 767.468433] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 767.468776] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 767.468974] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 767.477866] nova-conductor[52523]: DEBUG nova.quota [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Getting quotas for project bd0fcd8949d44cad8e75a3878331c428. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 767.480191] nova-conductor[52523]: DEBUG nova.quota [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Getting quotas for user e47f8eca093a4d5a82614eebe2f3a214 and project bd0fcd8949d44cad8e75a3878331c428. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 767.485603] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] [instance: fcf47169-eb7a-4644-bf3f-7150c44c247f] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 767.486051] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 767.486259] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 767.486429] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 767.490075] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] [instance: fcf47169-eb7a-4644-bf3f-7150c44c247f] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 767.490650] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 767.490856] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 767.491038] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 767.507379] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 767.507602] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 767.507774] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 803.673494] nova-conductor[52522]: ERROR nova.scheduler.utils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: 7a4778b7-5ffc-4641-b968-d0304fd67ee0] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 7a4778b7-5ffc-4641-b968-d0304fd67ee0 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 803.674099] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Rescheduling: True {{(pid=52522) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 803.674348] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 7a4778b7-5ffc-4641-b968-d0304fd67ee0.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 7a4778b7-5ffc-4641-b968-d0304fd67ee0. [ 803.674679] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: 7a4778b7-5ffc-4641-b968-d0304fd67ee0] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 7a4778b7-5ffc-4641-b968-d0304fd67ee0. [ 803.706959] nova-conductor[52522]: DEBUG nova.network.neutron [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: 7a4778b7-5ffc-4641-b968-d0304fd67ee0] deallocate_for_instance() {{(pid=52522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 803.733114] nova-conductor[52522]: DEBUG nova.network.neutron [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: 7a4778b7-5ffc-4641-b968-d0304fd67ee0] Instance cache missing network info. {{(pid=52522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.733793] nova-conductor[52522]: DEBUG nova.network.neutron [None req-70d4942f-629a-46a2-bfe7-7decf2e3298a tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: 7a4778b7-5ffc-4641-b968-d0304fd67ee0] Updating instance_info_cache with network_info: [] {{(pid=52522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.516797] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Took 0.11 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 804.529234] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 804.529468] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 804.529644] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 804.559688] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 804.559910] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 804.560095] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 804.560444] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 804.560628] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 804.560787] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 804.570300] nova-conductor[52523]: DEBUG nova.quota [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Getting quotas for project f0031e355e57421a8d48003a7eb717db. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 804.573262] nova-conductor[52523]: DEBUG nova.quota [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Getting quotas for user c06f3b2e0bd4459696b6724fa90f3809 and project f0031e355e57421a8d48003a7eb717db. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 804.578904] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: d55ee9a1-6921-4648-ace2-f2da13c3523e] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 804.579404] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 804.579714] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 804.579888] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 804.582974] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] [instance: d55ee9a1-6921-4648-ace2-f2da13c3523e] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 804.583628] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 804.583827] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 804.583992] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 804.598801] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 804.598989] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 804.599163] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 850.871268] nova-conductor[52523]: ERROR nova.scheduler.utils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] [instance: 363f5261-d589-4f99-b7dd-ab8f16cefee3] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 363f5261-d589-4f99-b7dd-ab8f16cefee3 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 850.872161] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Rescheduling: True {{(pid=52523) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 850.872486] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 363f5261-d589-4f99-b7dd-ab8f16cefee3.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 363f5261-d589-4f99-b7dd-ab8f16cefee3. [ 850.872924] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] [instance: 363f5261-d589-4f99-b7dd-ab8f16cefee3] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 363f5261-d589-4f99-b7dd-ab8f16cefee3. [ 850.902682] nova-conductor[52523]: DEBUG nova.network.neutron [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] [instance: 363f5261-d589-4f99-b7dd-ab8f16cefee3] deallocate_for_instance() {{(pid=52523) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.923283] nova-conductor[52523]: DEBUG nova.network.neutron [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] [instance: 363f5261-d589-4f99-b7dd-ab8f16cefee3] Instance cache missing network info. {{(pid=52523) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.927082] nova-conductor[52523]: DEBUG nova.network.neutron [None req-fac122e9-f2b2-46f3-ac71-9190de3e53b3 tempest-ServerExternalEventsTest-81672727 tempest-ServerExternalEventsTest-81672727-project-member] [instance: 363f5261-d589-4f99-b7dd-ab8f16cefee3] Updating instance_info_cache with network_info: [] {{(pid=52523) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 860.399365] nova-conductor[52523]: Traceback (most recent call last): [ 860.399365] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 860.399365] nova-conductor[52523]: return func(*args, **kwargs) [ 860.399365] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 860.399365] nova-conductor[52523]: selections = self._select_destinations( [ 860.399365] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 860.399365] nova-conductor[52523]: selections = self._schedule( [ 860.399365] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 860.399365] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 860.399365] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 860.399365] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 860.399365] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.399365] nova-conductor[52523]: ERROR nova.conductor.manager [ 860.407858] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 860.408100] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 860.408274] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 860.499786] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] [instance: 1e81817f-25ed-4912-9955-417ff3b3534c] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 860.500542] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 860.500814] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 860.500990] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 860.504344] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 860.504344] nova-conductor[52523]: Traceback (most recent call last): [ 860.504344] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 860.504344] nova-conductor[52523]: return func(*args, **kwargs) [ 860.504344] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 860.504344] nova-conductor[52523]: selections = self._select_destinations( [ 860.504344] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 860.504344] nova-conductor[52523]: selections = self._schedule( [ 860.504344] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 860.504344] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 860.504344] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 860.504344] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 860.504344] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 860.504344] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 860.505072] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-5b89dee5-3f8a-4fc4-8c67-067afc90d245 tempest-ServersTestJSON-2086144925 tempest-ServersTestJSON-2086144925-project-member] [instance: 1e81817f-25ed-4912-9955-417ff3b3534c] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 861.274563] nova-conductor[52522]: Traceback (most recent call last): [ 861.274563] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 861.274563] nova-conductor[52522]: return func(*args, **kwargs) [ 861.274563] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 861.274563] nova-conductor[52522]: selections = self._select_destinations( [ 861.274563] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 861.274563] nova-conductor[52522]: selections = self._schedule( [ 861.274563] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 861.274563] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 861.274563] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 861.274563] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 861.274563] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.274563] nova-conductor[52522]: ERROR nova.conductor.manager [ 861.281675] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 861.281960] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 861.282191] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 861.335463] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] [instance: ccb2ceab-76d4-4f74-99e3-cf235f03bc2b] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 861.336245] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 861.336615] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 861.336828] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 861.341754] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 861.341754] nova-conductor[52522]: Traceback (most recent call last): [ 861.341754] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 861.341754] nova-conductor[52522]: return func(*args, **kwargs) [ 861.341754] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 861.341754] nova-conductor[52522]: selections = self._select_destinations( [ 861.341754] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 861.341754] nova-conductor[52522]: selections = self._schedule( [ 861.341754] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 861.341754] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 861.341754] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 861.341754] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 861.341754] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 861.341754] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 861.342319] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-d7f195cc-845a-47d5-a78f-fbd2effe5d9d tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] [instance: ccb2ceab-76d4-4f74-99e3-cf235f03bc2b] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 862.295802] nova-conductor[52523]: Traceback (most recent call last): [ 862.295802] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 862.295802] nova-conductor[52523]: return func(*args, **kwargs) [ 862.295802] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 862.295802] nova-conductor[52523]: selections = self._select_destinations( [ 862.295802] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 862.295802] nova-conductor[52523]: selections = self._schedule( [ 862.295802] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 862.295802] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 862.295802] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 862.295802] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 862.295802] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.295802] nova-conductor[52523]: ERROR nova.conductor.manager [ 862.304278] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 862.304514] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 862.304723] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 862.361714] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 20aec966-1f49-4466-abdb-75656385d871] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 862.362504] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 862.362712] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 862.362836] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 862.366163] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 862.366163] nova-conductor[52523]: Traceback (most recent call last): [ 862.366163] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 862.366163] nova-conductor[52523]: return func(*args, **kwargs) [ 862.366163] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 862.366163] nova-conductor[52523]: selections = self._select_destinations( [ 862.366163] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 862.366163] nova-conductor[52523]: selections = self._schedule( [ 862.366163] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 862.366163] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 862.366163] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 862.366163] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 862.366163] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 862.366163] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 862.366742] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-fa6ec8aa-f1a3-4a2e-9578-7a8517ca94f6 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: 20aec966-1f49-4466-abdb-75656385d871] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 863.032632] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 863.032916] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 863.033083] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 863.126030] nova-conductor[52522]: Traceback (most recent call last): [ 863.126030] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 863.126030] nova-conductor[52522]: return func(*args, **kwargs) [ 863.126030] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 863.126030] nova-conductor[52522]: selections = self._select_destinations( [ 863.126030] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 863.126030] nova-conductor[52522]: selections = self._schedule( [ 863.126030] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 863.126030] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 863.126030] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 863.126030] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 863.126030] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.126030] nova-conductor[52522]: ERROR nova.conductor.manager [ 863.133341] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 863.133602] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 863.133968] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 863.209018] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] [instance: 3eb84b7a-2b36-44e2-8b64-4c6f0e950eb0] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 863.209158] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 863.209329] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 863.209494] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 863.212625] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 863.212625] nova-conductor[52522]: Traceback (most recent call last): [ 863.212625] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 863.212625] nova-conductor[52522]: return func(*args, **kwargs) [ 863.212625] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 863.212625] nova-conductor[52522]: selections = self._select_destinations( [ 863.212625] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 863.212625] nova-conductor[52522]: selections = self._schedule( [ 863.212625] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 863.212625] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 863.212625] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 863.212625] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 863.212625] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 863.212625] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 863.213471] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-6f9b51d6-7b35-4654-9c0e-b7ddbb7829d8 tempest-ServerGroupTestJSON-1405643382 tempest-ServerGroupTestJSON-1405643382-project-member] [instance: 3eb84b7a-2b36-44e2-8b64-4c6f0e950eb0] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.220419] nova-conductor[52523]: Traceback (most recent call last): [ 864.220419] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 864.220419] nova-conductor[52523]: return func(*args, **kwargs) [ 864.220419] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 864.220419] nova-conductor[52523]: selections = self._select_destinations( [ 864.220419] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 864.220419] nova-conductor[52523]: selections = self._schedule( [ 864.220419] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 864.220419] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 864.220419] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 864.220419] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 864.220419] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.220419] nova-conductor[52523]: ERROR nova.conductor.manager [ 864.227173] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 864.227500] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 864.227726] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 864.270260] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 48751b44-98ea-4f1c-b7e9-cfd92abb4cbf] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 864.271312] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 864.271312] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 864.271484] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 864.274651] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 864.274651] nova-conductor[52523]: Traceback (most recent call last): [ 864.274651] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 864.274651] nova-conductor[52523]: return func(*args, **kwargs) [ 864.274651] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 864.274651] nova-conductor[52523]: selections = self._select_destinations( [ 864.274651] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 864.274651] nova-conductor[52523]: selections = self._schedule( [ 864.274651] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 864.274651] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 864.274651] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 864.274651] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 864.274651] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 864.274651] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.275312] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-00bf6d64-30f7-4e48-ba61-b3873b6a45b2 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 48751b44-98ea-4f1c-b7e9-cfd92abb4cbf] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.559556] nova-conductor[52522]: Traceback (most recent call last): [ 864.559556] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 864.559556] nova-conductor[52522]: return func(*args, **kwargs) [ 864.559556] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 864.559556] nova-conductor[52522]: selections = self._select_destinations( [ 864.559556] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 864.559556] nova-conductor[52522]: selections = self._schedule( [ 864.559556] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 864.559556] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 864.559556] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 864.559556] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 864.559556] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.559556] nova-conductor[52522]: ERROR nova.conductor.manager [ 864.566064] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 864.566308] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 864.570071] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 864.617431] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: dcf89985-02e3-4bc6-9d7b-78247d495075] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 864.618059] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 864.618280] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 864.618450] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 864.622965] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 864.622965] nova-conductor[52522]: Traceback (most recent call last): [ 864.622965] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 864.622965] nova-conductor[52522]: return func(*args, **kwargs) [ 864.622965] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 864.622965] nova-conductor[52522]: selections = self._select_destinations( [ 864.622965] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 864.622965] nova-conductor[52522]: selections = self._schedule( [ 864.622965] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 864.622965] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 864.622965] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 864.622965] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 864.622965] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 864.622965] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 864.623356] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c9dcc6fb-7752-48ee-b3f1-ca675dc86b9f tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] [instance: dcf89985-02e3-4bc6-9d7b-78247d495075] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 865.838533] nova-conductor[52523]: Traceback (most recent call last): [ 865.838533] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 865.838533] nova-conductor[52523]: return func(*args, **kwargs) [ 865.838533] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 865.838533] nova-conductor[52523]: selections = self._select_destinations( [ 865.838533] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 865.838533] nova-conductor[52523]: selections = self._schedule( [ 865.838533] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 865.838533] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 865.838533] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 865.838533] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 865.838533] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.838533] nova-conductor[52523]: ERROR nova.conductor.manager [ 865.846445] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 865.846681] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 865.846857] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 865.891429] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 99c8658e-854b-4282-baea-63649a616238] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 865.892143] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 865.892336] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 865.892509] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 865.895547] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 865.895547] nova-conductor[52523]: Traceback (most recent call last): [ 865.895547] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 865.895547] nova-conductor[52523]: return func(*args, **kwargs) [ 865.895547] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 865.895547] nova-conductor[52523]: selections = self._select_destinations( [ 865.895547] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 865.895547] nova-conductor[52523]: selections = self._schedule( [ 865.895547] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 865.895547] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 865.895547] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 865.895547] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 865.895547] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 865.895547] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 865.896078] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-7ceb71f8-d975-4410-951e-5a0fb6b1aa3c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 99c8658e-854b-4282-baea-63649a616238] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 866.061417] nova-conductor[52522]: Traceback (most recent call last): [ 866.061417] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 866.061417] nova-conductor[52522]: return func(*args, **kwargs) [ 866.061417] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 866.061417] nova-conductor[52522]: selections = self._select_destinations( [ 866.061417] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 866.061417] nova-conductor[52522]: selections = self._schedule( [ 866.061417] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 866.061417] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 866.061417] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 866.061417] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 866.061417] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.061417] nova-conductor[52522]: ERROR nova.conductor.manager [ 866.069536] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 866.069865] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 866.070119] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 866.110697] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] [instance: ed59b3c8-7da9-43a3-b61d-3e4e784b3227] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 866.111714] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 866.111881] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 866.112130] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 866.119118] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 866.119118] nova-conductor[52522]: Traceback (most recent call last): [ 866.119118] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 866.119118] nova-conductor[52522]: return func(*args, **kwargs) [ 866.119118] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 866.119118] nova-conductor[52522]: selections = self._select_destinations( [ 866.119118] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 866.119118] nova-conductor[52522]: selections = self._schedule( [ 866.119118] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 866.119118] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 866.119118] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 866.119118] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 866.119118] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 866.119118] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 866.119118] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-b6e7d133-b3e7-43a9-a18b-4d350dc4fc11 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] [instance: ed59b3c8-7da9-43a3-b61d-3e4e784b3227] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 871.188604] nova-conductor[52523]: Traceback (most recent call last): [ 871.188604] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 871.188604] nova-conductor[52523]: return func(*args, **kwargs) [ 871.188604] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 871.188604] nova-conductor[52523]: selections = self._select_destinations( [ 871.188604] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 871.188604] nova-conductor[52523]: selections = self._schedule( [ 871.188604] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 871.188604] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 871.188604] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 871.188604] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 871.188604] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.188604] nova-conductor[52523]: ERROR nova.conductor.manager [ 871.196446] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 871.196702] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 871.196880] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 871.240078] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 2f7ff509-b981-4502-a2ab-7f69fc70a9d5] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 871.240874] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 871.243021] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 871.243021] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 871.247469] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 871.247469] nova-conductor[52523]: Traceback (most recent call last): [ 871.247469] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 871.247469] nova-conductor[52523]: return func(*args, **kwargs) [ 871.247469] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 871.247469] nova-conductor[52523]: selections = self._select_destinations( [ 871.247469] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 871.247469] nova-conductor[52523]: selections = self._schedule( [ 871.247469] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 871.247469] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 871.247469] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 871.247469] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 871.247469] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 871.247469] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 871.248774] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-caa665b4-6f6f-4c26-afd2-c2e2273c37cc tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 2f7ff509-b981-4502-a2ab-7f69fc70a9d5] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 872.297078] nova-conductor[52522]: Traceback (most recent call last): [ 872.297078] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 872.297078] nova-conductor[52522]: return func(*args, **kwargs) [ 872.297078] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 872.297078] nova-conductor[52522]: selections = self._select_destinations( [ 872.297078] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 872.297078] nova-conductor[52522]: selections = self._schedule( [ 872.297078] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 872.297078] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 872.297078] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 872.297078] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 872.297078] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.297078] nova-conductor[52522]: ERROR nova.conductor.manager [ 872.307835] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 872.308093] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 872.308266] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 872.372442] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] [instance: 65265d5f-f76f-477b-83e2-b48fe05b6655] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 872.372754] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 872.373028] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 872.373206] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 872.379753] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 872.379753] nova-conductor[52522]: Traceback (most recent call last): [ 872.379753] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 872.379753] nova-conductor[52522]: return func(*args, **kwargs) [ 872.379753] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 872.379753] nova-conductor[52522]: selections = self._select_destinations( [ 872.379753] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 872.379753] nova-conductor[52522]: selections = self._schedule( [ 872.379753] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 872.379753] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 872.379753] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 872.379753] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 872.379753] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 872.379753] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 872.380466] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-f2be6009-8305-4483-be87-6fc424d4b454 tempest-ServerActionsTestOtherB-814910926 tempest-ServerActionsTestOtherB-814910926-project-member] [instance: 65265d5f-f76f-477b-83e2-b48fe05b6655] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.494635] nova-conductor[52523]: Traceback (most recent call last): [ 873.494635] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 873.494635] nova-conductor[52523]: return func(*args, **kwargs) [ 873.494635] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 873.494635] nova-conductor[52523]: selections = self._select_destinations( [ 873.494635] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 873.494635] nova-conductor[52523]: selections = self._schedule( [ 873.494635] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 873.494635] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 873.494635] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 873.494635] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 873.494635] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.494635] nova-conductor[52523]: ERROR nova.conductor.manager [ 873.501251] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 873.501484] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 873.501660] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 873.552489] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] [instance: 80f3c623-fdd3-4410-869d-8cf3f54b02ef] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 873.553252] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 873.553483] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 873.553635] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 873.557719] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 873.557719] nova-conductor[52523]: Traceback (most recent call last): [ 873.557719] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 873.557719] nova-conductor[52523]: return func(*args, **kwargs) [ 873.557719] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 873.557719] nova-conductor[52523]: selections = self._select_destinations( [ 873.557719] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 873.557719] nova-conductor[52523]: selections = self._schedule( [ 873.557719] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 873.557719] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 873.557719] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 873.557719] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 873.557719] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 873.557719] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.558315] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-9c95e371-d886-46e0-9a1c-3471b081ef86 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] [instance: 80f3c623-fdd3-4410-869d-8cf3f54b02ef] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.764665] nova-conductor[52522]: Traceback (most recent call last): [ 873.764665] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 873.764665] nova-conductor[52522]: return func(*args, **kwargs) [ 873.764665] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 873.764665] nova-conductor[52522]: selections = self._select_destinations( [ 873.764665] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 873.764665] nova-conductor[52522]: selections = self._schedule( [ 873.764665] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 873.764665] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 873.764665] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 873.764665] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 873.764665] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.764665] nova-conductor[52522]: ERROR nova.conductor.manager [ 873.773461] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 873.773651] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 873.773837] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 873.828835] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 3dd852dd-aaf7-4e35-bf83-0e34aa553c3c] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 873.830283] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 873.830501] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 873.830683] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 873.834655] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 873.834655] nova-conductor[52522]: Traceback (most recent call last): [ 873.834655] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 873.834655] nova-conductor[52522]: return func(*args, **kwargs) [ 873.834655] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 873.834655] nova-conductor[52522]: selections = self._select_destinations( [ 873.834655] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 873.834655] nova-conductor[52522]: selections = self._schedule( [ 873.834655] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 873.834655] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 873.834655] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 873.834655] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 873.834655] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 873.834655] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 873.835468] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-986bf0ea-55c8-49bd-842a-d6d0ba6b1929 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 3dd852dd-aaf7-4e35-bf83-0e34aa553c3c] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 874.281508] nova-conductor[52523]: Traceback (most recent call last): [ 874.281508] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 874.281508] nova-conductor[52523]: return func(*args, **kwargs) [ 874.281508] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 874.281508] nova-conductor[52523]: selections = self._select_destinations( [ 874.281508] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 874.281508] nova-conductor[52523]: selections = self._schedule( [ 874.281508] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 874.281508] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 874.281508] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 874.281508] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 874.281508] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.281508] nova-conductor[52523]: ERROR nova.conductor.manager [ 874.290531] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 874.291019] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 874.291109] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 874.349702] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] [instance: 7b000455-28ca-48b2-b736-92f41fc89efe] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 874.349702] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 874.349702] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 874.349702] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 874.351996] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 874.351996] nova-conductor[52523]: Traceback (most recent call last): [ 874.351996] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 874.351996] nova-conductor[52523]: return func(*args, **kwargs) [ 874.351996] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 874.351996] nova-conductor[52523]: selections = self._select_destinations( [ 874.351996] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 874.351996] nova-conductor[52523]: selections = self._schedule( [ 874.351996] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 874.351996] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 874.351996] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 874.351996] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 874.351996] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 874.351996] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 874.353375] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-9d5750b6-4079-4ae5-934c-76a3922c16f2 tempest-ServerShowV247Test-733303285 tempest-ServerShowV247Test-733303285-project-member] [instance: 7b000455-28ca-48b2-b736-92f41fc89efe] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 876.380144] nova-conductor[52523]: Traceback (most recent call last): [ 876.380144] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 876.380144] nova-conductor[52523]: return func(*args, **kwargs) [ 876.380144] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 876.380144] nova-conductor[52523]: selections = self._select_destinations( [ 876.380144] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 876.380144] nova-conductor[52523]: selections = self._schedule( [ 876.380144] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 876.380144] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 876.380144] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 876.380144] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 876.380144] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.380144] nova-conductor[52523]: ERROR nova.conductor.manager [ 876.388473] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 876.388799] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 876.388882] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 876.438204] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: cac1dc1a-7b98-4390-a928-45617d9a8ab3] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 876.438970] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 876.439211] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 876.439428] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 876.450681] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 876.450681] nova-conductor[52523]: Traceback (most recent call last): [ 876.450681] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 876.450681] nova-conductor[52523]: return func(*args, **kwargs) [ 876.450681] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 876.450681] nova-conductor[52523]: selections = self._select_destinations( [ 876.450681] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 876.450681] nova-conductor[52523]: selections = self._schedule( [ 876.450681] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 876.450681] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 876.450681] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 876.450681] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 876.450681] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 876.450681] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 876.452222] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-15106646-0fbb-4219-a77b-06677c4a537c tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: cac1dc1a-7b98-4390-a928-45617d9a8ab3] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 878.760429] nova-conductor[52522]: Traceback (most recent call last): [ 878.760429] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 878.760429] nova-conductor[52522]: return func(*args, **kwargs) [ 878.760429] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 878.760429] nova-conductor[52522]: selections = self._select_destinations( [ 878.760429] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 878.760429] nova-conductor[52522]: selections = self._schedule( [ 878.760429] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 878.760429] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 878.760429] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 878.760429] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 878.760429] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.760429] nova-conductor[52522]: ERROR nova.conductor.manager [ 878.768881] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 878.769137] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 878.769921] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 878.832561] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 7ce4c7be-193e-442f-bbc4-1b7005d43921] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 878.834546] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 878.834546] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 878.834546] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 878.849434] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 878.849434] nova-conductor[52522]: Traceback (most recent call last): [ 878.849434] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 878.849434] nova-conductor[52522]: return func(*args, **kwargs) [ 878.849434] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 878.849434] nova-conductor[52522]: selections = self._select_destinations( [ 878.849434] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 878.849434] nova-conductor[52522]: selections = self._schedule( [ 878.849434] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 878.849434] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 878.849434] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 878.849434] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 878.849434] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 878.849434] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 878.850079] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-3a254be9-56b0-4c24-8cfd-b3def49ef17e tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] [instance: 7ce4c7be-193e-442f-bbc4-1b7005d43921] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 879.861324] nova-conductor[52523]: Traceback (most recent call last): [ 879.861324] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 879.861324] nova-conductor[52523]: return func(*args, **kwargs) [ 879.861324] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 879.861324] nova-conductor[52523]: selections = self._select_destinations( [ 879.861324] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 879.861324] nova-conductor[52523]: selections = self._schedule( [ 879.861324] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 879.861324] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 879.861324] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 879.861324] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 879.861324] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.861324] nova-conductor[52523]: ERROR nova.conductor.manager [ 879.879328] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 879.879328] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.004s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 879.879328] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 879.935939] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] [instance: c61acd23-9e6c-4a18-9b18-b19408850cbe] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 879.937649] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 879.937649] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 879.937649] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 879.940531] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 879.940531] nova-conductor[52523]: Traceback (most recent call last): [ 879.940531] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 879.940531] nova-conductor[52523]: return func(*args, **kwargs) [ 879.940531] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 879.940531] nova-conductor[52523]: selections = self._select_destinations( [ 879.940531] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 879.940531] nova-conductor[52523]: selections = self._schedule( [ 879.940531] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 879.940531] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 879.940531] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 879.940531] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 879.940531] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 879.940531] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 879.941096] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-64503bec-43ae-4d52-ae2d-b7e2d10bc384 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] [instance: c61acd23-9e6c-4a18-9b18-b19408850cbe] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 883.802796] nova-conductor[52523]: Traceback (most recent call last): [ 883.802796] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 883.802796] nova-conductor[52523]: return func(*args, **kwargs) [ 883.802796] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 883.802796] nova-conductor[52523]: selections = self._select_destinations( [ 883.802796] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 883.802796] nova-conductor[52523]: selections = self._schedule( [ 883.802796] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 883.802796] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 883.802796] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 883.802796] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 883.802796] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.802796] nova-conductor[52523]: ERROR nova.conductor.manager [ 883.813868] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 883.814174] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 883.814341] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 883.863701] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] [instance: a04b1b2f-26ce-480b-b786-70e4d689eceb] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 883.864609] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 883.865345] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 883.865593] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 883.871650] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 883.871650] nova-conductor[52523]: Traceback (most recent call last): [ 883.871650] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 883.871650] nova-conductor[52523]: return func(*args, **kwargs) [ 883.871650] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 883.871650] nova-conductor[52523]: selections = self._select_destinations( [ 883.871650] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 883.871650] nova-conductor[52523]: selections = self._schedule( [ 883.871650] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 883.871650] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 883.871650] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 883.871650] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 883.871650] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 883.871650] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 883.872486] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-06e92b60-b210-4c0f-b58b-2e69270ff1a9 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] [instance: a04b1b2f-26ce-480b-b786-70e4d689eceb] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 884.570227] nova-conductor[52522]: Traceback (most recent call last): [ 884.570227] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 884.570227] nova-conductor[52522]: return func(*args, **kwargs) [ 884.570227] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 884.570227] nova-conductor[52522]: selections = self._select_destinations( [ 884.570227] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 884.570227] nova-conductor[52522]: selections = self._schedule( [ 884.570227] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 884.570227] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 884.570227] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 884.570227] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 884.570227] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.570227] nova-conductor[52522]: ERROR nova.conductor.manager [ 884.576204] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 884.576475] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 884.576678] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 884.634044] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] [instance: 3586f01e-3213-4a6a-9147-f3fa1e738e03] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 884.634887] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 884.635123] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 884.635333] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 884.642372] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 884.642372] nova-conductor[52522]: Traceback (most recent call last): [ 884.642372] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 884.642372] nova-conductor[52522]: return func(*args, **kwargs) [ 884.642372] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 884.642372] nova-conductor[52522]: selections = self._select_destinations( [ 884.642372] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 884.642372] nova-conductor[52522]: selections = self._schedule( [ 884.642372] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 884.642372] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 884.642372] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 884.642372] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 884.642372] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 884.642372] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 884.644211] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-668c02f4-450f-4380-b62b-9fa13097ba98 tempest-ServerShowV254Test-1431718957 tempest-ServerShowV254Test-1431718957-project-member] [instance: 3586f01e-3213-4a6a-9147-f3fa1e738e03] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 885.552885] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Took 0.16 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 885.569496] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 885.569496] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 885.569496] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 885.633016] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 885.633338] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 885.633553] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 885.633966] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 885.634267] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 885.634473] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 885.643974] nova-conductor[52522]: DEBUG nova.quota [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Getting quotas for project 60bf6fde086045b492b838eab8435479. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 885.647108] nova-conductor[52522]: DEBUG nova.quota [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Getting quotas for user b0100cdbf36b40da85334a72b9121fc5 and project 60bf6fde086045b492b838eab8435479. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 885.653296] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] [instance: e8ed8e30-1b1d-4ab4-abd7-e68cb72916cb] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 885.653796] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 885.654059] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 885.654273] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 885.659204] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] [instance: e8ed8e30-1b1d-4ab4-abd7-e68cb72916cb] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 885.659982] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 885.660136] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 885.660316] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 885.673376] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 885.673600] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 885.673774] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 887.596011] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Took 0.12 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 887.607621] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 887.607788] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 887.607960] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 887.637991] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 887.638727] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 887.638727] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 887.638727] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 887.638980] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 887.639054] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 887.651532] nova-conductor[52523]: DEBUG nova.quota [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Getting quotas for project 9fb45b1c3ac3465c8b7737a90a8a968e. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 887.653980] nova-conductor[52523]: DEBUG nova.quota [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Getting quotas for user 885cac70fdc44efaaa3b2f0e1bfb2907 and project 9fb45b1c3ac3465c8b7737a90a8a968e. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 887.659622] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] [instance: 54ded864-1c3e-4a47-968f-ca597c82cb87] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 887.660230] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 887.660434] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 887.660600] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 887.663739] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] [instance: 54ded864-1c3e-4a47-968f-ca597c82cb87] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 887.664551] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 887.664750] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 887.664913] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 887.677820] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 887.678039] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 887.678232] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 897.593067] nova-conductor[52523]: ERROR nova.scheduler.utils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: ce222a29-3611-45b3-9664-87ae2fb1b1b8] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance ce222a29-3611-45b3-9664-87ae2fb1b1b8 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 897.593067] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Rescheduling: True {{(pid=52523) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 897.593067] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance ce222a29-3611-45b3-9664-87ae2fb1b1b8.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance ce222a29-3611-45b3-9664-87ae2fb1b1b8. [ 897.593067] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: ce222a29-3611-45b3-9664-87ae2fb1b1b8] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance ce222a29-3611-45b3-9664-87ae2fb1b1b8. [ 897.616746] nova-conductor[52523]: DEBUG nova.network.neutron [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: ce222a29-3611-45b3-9664-87ae2fb1b1b8] deallocate_for_instance() {{(pid=52523) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 897.631585] nova-conductor[52522]: DEBUG nova.db.main.api [None req-5269542d-1bd0-4a60-91e5-e702adcd1908 tempest-AttachInterfacesTestJSON-62544092 tempest-AttachInterfacesTestJSON-62544092-project-member] Created instance_extra for 1eaf8e02-bfb0-4928-9687-cc781a84d16d {{(pid=52522) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 897.636661] nova-conductor[52523]: DEBUG nova.network.neutron [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: ce222a29-3611-45b3-9664-87ae2fb1b1b8] Instance cache missing network info. {{(pid=52523) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 897.640084] nova-conductor[52523]: DEBUG nova.network.neutron [None req-3989da87-6fee-44a7-bc2b-45950fa18c6f tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: ce222a29-3611-45b3-9664-87ae2fb1b1b8] Updating instance_info_cache with network_info: [] {{(pid=52523) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.691424] nova-conductor[52523]: DEBUG nova.db.main.api [None req-14455799-885d-47a6-b31a-d48adadc4279 tempest-ImagesTestJSON-1812607055 tempest-ImagesTestJSON-1812607055-project-member] Created instance_extra for c841298b-f103-4dc7-8884-efdf2ebc20a6 {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 897.763822] nova-conductor[52522]: DEBUG nova.db.main.api [None req-2a575c2b-5102-4501-a724-cecb4fb9a882 tempest-ServersTestJSON-1695110591 tempest-ServersTestJSON-1695110591-project-member] Created instance_extra for bf70d23b-4ab5-476e-814c-264b6a9f2455 {{(pid=52522) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 897.822435] nova-conductor[52523]: DEBUG nova.db.main.api [None req-b87ef477-9e92-4961-845a-6acd6bee3a06 tempest-ServersNegativeTestJSON-1972359588 tempest-ServersNegativeTestJSON-1972359588-project-member] Created instance_extra for 67ff7d52-6e30-4730-9b5a-9ae32f68b953 {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 897.877830] nova-conductor[52522]: DEBUG nova.db.main.api [None req-ba27ce61-be1b-476b-9184-384249a73295 tempest-SecurityGroupsTestJSON-2088684547 tempest-SecurityGroupsTestJSON-2088684547-project-member] Created instance_extra for bd91e947-acae-4dbd-b48b-5a6727eb4cbb {{(pid=52522) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 897.931009] nova-conductor[52523]: DEBUG nova.db.main.api [None req-798944be-b318-4a6d-a582-3984731e10fc tempest-ServerActionsTestOtherA-202760684 tempest-ServerActionsTestOtherA-202760684-project-member] Created instance_extra for 95358801-c9d8-4582-a712-36a8bf586456 {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 897.986018] nova-conductor[52522]: DEBUG nova.db.main.api [None req-814fad23-a147-41ec-ba07-f0ec4ac3c42f tempest-ServerActionsV293TestJSON-332701173 tempest-ServerActionsV293TestJSON-332701173-project-member] Created instance_extra for be6a4290-dbb3-4e1f-bdd4-0dc106db9435 {{(pid=52522) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 900.867476] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Took 0.11 seconds to select destinations for 1 instance(s). {{(pid=52523) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 900.880126] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 900.880126] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 900.880126] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 900.908703] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 900.909300] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 900.910565] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 900.910565] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 900.910565] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 900.910565] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 900.920025] nova-conductor[52523]: DEBUG nova.quota [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Getting quotas for project 50f58441c5ba4f09a505f350adf21708. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 900.922045] nova-conductor[52523]: DEBUG nova.quota [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Getting quotas for user 46b1172c3120415cb62749437bbff1ce and project 50f58441c5ba4f09a505f350adf21708. Resources: {'cores', 'instances', 'ram'} {{(pid=52523) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 900.928025] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] [instance: e259637a-0fc8-4368-8a7a-c15a134ed17d] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52523) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 900.928731] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 900.929094] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 900.930052] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 900.932381] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] [instance: e259637a-0fc8-4368-8a7a-c15a134ed17d] block_device_mapping [BlockDeviceMapping(attachment_id=22944502-06d8-4caa-9d68-ae0dd4b27001,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='volume',device_name=None,device_type=None,disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id=None,instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='volume',tag=None,updated_at=,uuid=,volume_id='a68c7f7d-bdca-4eb3-98b9-e7d6e3a78129',volume_size=1,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 900.934572] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 900.934572] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 900.934572] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 900.948631] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 900.949213] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 900.951117] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 948.496824] nova-conductor[52523]: ERROR nova.scheduler.utils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] [instance: 9214a18f-c22d-4e24-980e-7241a2b993bd] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 9214a18f-c22d-4e24-980e-7241a2b993bd was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 948.497504] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Rescheduling: True {{(pid=52523) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 948.497774] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 9214a18f-c22d-4e24-980e-7241a2b993bd.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 9214a18f-c22d-4e24-980e-7241a2b993bd. [ 948.498166] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] [instance: 9214a18f-c22d-4e24-980e-7241a2b993bd] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 9214a18f-c22d-4e24-980e-7241a2b993bd. [ 948.523008] nova-conductor[52523]: DEBUG nova.network.neutron [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] [instance: 9214a18f-c22d-4e24-980e-7241a2b993bd] deallocate_for_instance() {{(pid=52523) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.540983] nova-conductor[52523]: DEBUG nova.network.neutron [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] [instance: 9214a18f-c22d-4e24-980e-7241a2b993bd] Instance cache missing network info. {{(pid=52523) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.544893] nova-conductor[52523]: DEBUG nova.network.neutron [None req-6a0399c7-51de-4d53-b231-40ed4691f878 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658 tempest-FloatingIPsAssociationNegativeTestJSON-1831931658-project-member] [instance: 9214a18f-c22d-4e24-980e-7241a2b993bd] Updating instance_info_cache with network_info: [] {{(pid=52523) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 955.084162] nova-conductor[52523]: Traceback (most recent call last): [ 955.084162] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 955.084162] nova-conductor[52523]: return func(*args, **kwargs) [ 955.084162] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 955.084162] nova-conductor[52523]: selections = self._select_destinations( [ 955.084162] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 955.084162] nova-conductor[52523]: selections = self._schedule( [ 955.084162] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 955.084162] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 955.084162] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 955.084162] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 955.084162] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.084162] nova-conductor[52523]: ERROR nova.conductor.manager [ 955.091169] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 955.091442] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 955.091620] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 955.133743] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 05e761d8-26ad-49c0-ab74-3ff6b13e7322] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 955.134486] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 955.134715] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 955.134889] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 955.137803] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 955.137803] nova-conductor[52523]: Traceback (most recent call last): [ 955.137803] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 955.137803] nova-conductor[52523]: return func(*args, **kwargs) [ 955.137803] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 955.137803] nova-conductor[52523]: selections = self._select_destinations( [ 955.137803] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 955.137803] nova-conductor[52523]: selections = self._schedule( [ 955.137803] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 955.137803] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 955.137803] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 955.137803] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 955.137803] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 955.137803] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 955.138402] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-431a23bd-5652-4d71-999e-c1b42b86476c tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 05e761d8-26ad-49c0-ab74-3ff6b13e7322] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 956.927228] nova-conductor[52522]: Traceback (most recent call last): [ 956.927228] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 956.927228] nova-conductor[52522]: return func(*args, **kwargs) [ 956.927228] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 956.927228] nova-conductor[52522]: selections = self._select_destinations( [ 956.927228] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 956.927228] nova-conductor[52522]: selections = self._schedule( [ 956.927228] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 956.927228] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 956.927228] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 956.927228] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 956.927228] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.927228] nova-conductor[52522]: ERROR nova.conductor.manager [ 956.934046] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 956.934345] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 956.934581] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 956.972349] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: aac61a06-f49e-4bb9-9832-d2d39ca05275] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 956.973065] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 956.973296] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 956.973506] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 956.976393] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 956.976393] nova-conductor[52522]: Traceback (most recent call last): [ 956.976393] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 956.976393] nova-conductor[52522]: return func(*args, **kwargs) [ 956.976393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 956.976393] nova-conductor[52522]: selections = self._select_destinations( [ 956.976393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 956.976393] nova-conductor[52522]: selections = self._schedule( [ 956.976393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 956.976393] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 956.976393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 956.976393] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 956.976393] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 956.976393] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 956.977036] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-aedf829b-dc23-4ef4-8c9e-afdccb3d0231 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: aac61a06-f49e-4bb9-9832-d2d39ca05275] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 958.620520] nova-conductor[52523]: Traceback (most recent call last): [ 958.620520] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 958.620520] nova-conductor[52523]: return func(*args, **kwargs) [ 958.620520] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 958.620520] nova-conductor[52523]: selections = self._select_destinations( [ 958.620520] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 958.620520] nova-conductor[52523]: selections = self._schedule( [ 958.620520] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 958.620520] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 958.620520] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 958.620520] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 958.620520] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.620520] nova-conductor[52523]: ERROR nova.conductor.manager [ 958.627506] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 958.627744] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 958.627981] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 958.668187] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 26da8dbe-de74-4e84-a643-3bb3a00236d2] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 958.668870] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 958.669099] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 958.669279] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 958.671989] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 958.671989] nova-conductor[52523]: Traceback (most recent call last): [ 958.671989] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 958.671989] nova-conductor[52523]: return func(*args, **kwargs) [ 958.671989] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 958.671989] nova-conductor[52523]: selections = self._select_destinations( [ 958.671989] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 958.671989] nova-conductor[52523]: selections = self._schedule( [ 958.671989] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 958.671989] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 958.671989] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 958.671989] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 958.671989] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 958.671989] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 958.672521] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-14b94c08-b1b5-4bb8-a2ae-137a0006daea tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 26da8dbe-de74-4e84-a643-3bb3a00236d2] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 960.398393] nova-conductor[52522]: Traceback (most recent call last): [ 960.398393] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 960.398393] nova-conductor[52522]: return func(*args, **kwargs) [ 960.398393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 960.398393] nova-conductor[52522]: selections = self._select_destinations( [ 960.398393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 960.398393] nova-conductor[52522]: selections = self._schedule( [ 960.398393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 960.398393] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 960.398393] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 960.398393] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 960.398393] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.398393] nova-conductor[52522]: ERROR nova.conductor.manager [ 960.404988] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 960.405241] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 960.405418] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 960.445722] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 5c2cca43-b6f3-4b5e-b612-be93280c7297] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 960.446374] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 960.446527] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 960.446697] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 960.450091] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 960.450091] nova-conductor[52522]: Traceback (most recent call last): [ 960.450091] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 960.450091] nova-conductor[52522]: return func(*args, **kwargs) [ 960.450091] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 960.450091] nova-conductor[52522]: selections = self._select_destinations( [ 960.450091] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 960.450091] nova-conductor[52522]: selections = self._schedule( [ 960.450091] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 960.450091] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 960.450091] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 960.450091] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 960.450091] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 960.450091] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 960.450621] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-ba4c4c23-5311-40d6-bb52-fcec332b95b7 tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 5c2cca43-b6f3-4b5e-b612-be93280c7297] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 962.445792] nova-conductor[52523]: Traceback (most recent call last): [ 962.445792] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 962.445792] nova-conductor[52523]: return func(*args, **kwargs) [ 962.445792] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 962.445792] nova-conductor[52523]: selections = self._select_destinations( [ 962.445792] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 962.445792] nova-conductor[52523]: selections = self._schedule( [ 962.445792] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 962.445792] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 962.445792] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 962.445792] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 962.445792] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.445792] nova-conductor[52523]: ERROR nova.conductor.manager [ 962.453469] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 962.453718] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 962.453893] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 962.493153] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 62656da6-449c-47cf-901a-baf5ffd06cfb] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 962.493902] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 962.494130] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 962.494303] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 962.497101] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 962.497101] nova-conductor[52523]: Traceback (most recent call last): [ 962.497101] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 962.497101] nova-conductor[52523]: return func(*args, **kwargs) [ 962.497101] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 962.497101] nova-conductor[52523]: selections = self._select_destinations( [ 962.497101] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 962.497101] nova-conductor[52523]: selections = self._schedule( [ 962.497101] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 962.497101] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 962.497101] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 962.497101] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 962.497101] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 962.497101] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 962.497619] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-31c2de6c-b95b-4361-995c-fbf6282e43bc tempest-ServerDiskConfigTestJSON-1882542127 tempest-ServerDiskConfigTestJSON-1882542127-project-member] [instance: 62656da6-449c-47cf-901a-baf5ffd06cfb] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 969.825216] nova-conductor[52522]: Traceback (most recent call last): [ 969.825216] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 969.825216] nova-conductor[52522]: return func(*args, **kwargs) [ 969.825216] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 969.825216] nova-conductor[52522]: selections = self._select_destinations( [ 969.825216] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 969.825216] nova-conductor[52522]: selections = self._schedule( [ 969.825216] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 969.825216] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 969.825216] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 969.825216] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 969.825216] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.825216] nova-conductor[52522]: ERROR nova.conductor.manager [ 969.837581] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 969.837581] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 969.837581] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 969.894658] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 178e7746-5dd2-4206-94ba-a6f3bf8532c3] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 969.897525] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 969.897525] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 969.897525] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 969.899864] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 969.899864] nova-conductor[52522]: Traceback (most recent call last): [ 969.899864] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 969.899864] nova-conductor[52522]: return func(*args, **kwargs) [ 969.899864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 969.899864] nova-conductor[52522]: selections = self._select_destinations( [ 969.899864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 969.899864] nova-conductor[52522]: selections = self._schedule( [ 969.899864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 969.899864] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 969.899864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 969.899864] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 969.899864] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 969.899864] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 969.900470] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-88f99e1e-7059-4f8f-989a-c9b73268fed0 tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 178e7746-5dd2-4206-94ba-a6f3bf8532c3] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 969.984322] nova-conductor[52523]: Traceback (most recent call last): [ 969.984322] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 969.984322] nova-conductor[52523]: return func(*args, **kwargs) [ 969.984322] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 969.984322] nova-conductor[52523]: selections = self._select_destinations( [ 969.984322] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 969.984322] nova-conductor[52523]: selections = self._schedule( [ 969.984322] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 969.984322] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 969.984322] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 969.984322] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 969.984322] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.984322] nova-conductor[52523]: ERROR nova.conductor.manager [ 969.992957] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 969.993238] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 969.993457] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 970.059356] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] [instance: bf278e90-1194-4df2-aa3a-b62157412c7f] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 970.062568] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 970.062568] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 970.062568] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 970.064217] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 970.064217] nova-conductor[52523]: Traceback (most recent call last): [ 970.064217] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 970.064217] nova-conductor[52523]: return func(*args, **kwargs) [ 970.064217] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 970.064217] nova-conductor[52523]: selections = self._select_destinations( [ 970.064217] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 970.064217] nova-conductor[52523]: selections = self._schedule( [ 970.064217] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 970.064217] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 970.064217] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 970.064217] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 970.064217] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 970.064217] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 970.064787] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-398d99fe-00c8-4369-a608-b1a2a85b15a7 tempest-InstanceActionsNegativeTestJSON-312224709 tempest-InstanceActionsNegativeTestJSON-312224709-project-member] [instance: bf278e90-1194-4df2-aa3a-b62157412c7f] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 970.966862] nova-conductor[52522]: Traceback (most recent call last): [ 970.966862] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 970.966862] nova-conductor[52522]: return func(*args, **kwargs) [ 970.966862] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 970.966862] nova-conductor[52522]: selections = self._select_destinations( [ 970.966862] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 970.966862] nova-conductor[52522]: selections = self._schedule( [ 970.966862] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 970.966862] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 970.966862] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 970.966862] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 970.966862] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.966862] nova-conductor[52522]: ERROR nova.conductor.manager [ 970.975568] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 970.975568] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 970.975568] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 971.044022] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 84fb37ba-4379-4164-b2d9-e58cd51767de] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 971.044022] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 971.044022] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 971.044022] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 971.047359] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 971.047359] nova-conductor[52522]: Traceback (most recent call last): [ 971.047359] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 971.047359] nova-conductor[52522]: return func(*args, **kwargs) [ 971.047359] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 971.047359] nova-conductor[52522]: selections = self._select_destinations( [ 971.047359] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 971.047359] nova-conductor[52522]: selections = self._schedule( [ 971.047359] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 971.047359] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 971.047359] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 971.047359] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 971.047359] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 971.047359] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 971.047898] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-4c4ffd6d-12a6-4d55-bf52-328fc3538395 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 84fb37ba-4379-4164-b2d9-e58cd51767de] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 973.132164] nova-conductor[52523]: Traceback (most recent call last): [ 973.132164] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 973.132164] nova-conductor[52523]: return func(*args, **kwargs) [ 973.132164] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 973.132164] nova-conductor[52523]: selections = self._select_destinations( [ 973.132164] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 973.132164] nova-conductor[52523]: selections = self._schedule( [ 973.132164] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 973.132164] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 973.132164] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 973.132164] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 973.132164] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.132164] nova-conductor[52523]: ERROR nova.conductor.manager [ 973.138864] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 973.139123] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 973.141062] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 973.187628] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 25710e7a-13fc-45a3-b342-3d8480b0d1bd] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 973.188390] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 973.188625] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 973.188769] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 973.192609] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 973.192609] nova-conductor[52523]: Traceback (most recent call last): [ 973.192609] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 973.192609] nova-conductor[52523]: return func(*args, **kwargs) [ 973.192609] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 973.192609] nova-conductor[52523]: selections = self._select_destinations( [ 973.192609] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 973.192609] nova-conductor[52523]: selections = self._schedule( [ 973.192609] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 973.192609] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 973.192609] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 973.192609] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 973.192609] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 973.192609] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 973.193139] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-70bb5e87-a082-4cf9-83bb-c0202753da6a tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 25710e7a-13fc-45a3-b342-3d8480b0d1bd] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 974.870813] nova-conductor[52522]: Traceback (most recent call last): [ 974.870813] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 974.870813] nova-conductor[52522]: return func(*args, **kwargs) [ 974.870813] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 974.870813] nova-conductor[52522]: selections = self._select_destinations( [ 974.870813] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 974.870813] nova-conductor[52522]: selections = self._schedule( [ 974.870813] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 974.870813] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 974.870813] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 974.870813] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 974.870813] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.870813] nova-conductor[52522]: ERROR nova.conductor.manager [ 974.877175] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 974.877406] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 974.877606] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 974.915664] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: ddeaa346-8d42-43ad-b4d5-d70bc6618c67] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 974.916815] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 974.917064] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 974.917263] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 974.920375] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 974.920375] nova-conductor[52522]: Traceback (most recent call last): [ 974.920375] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 974.920375] nova-conductor[52522]: return func(*args, **kwargs) [ 974.920375] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 974.920375] nova-conductor[52522]: selections = self._select_destinations( [ 974.920375] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 974.920375] nova-conductor[52522]: selections = self._schedule( [ 974.920375] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 974.920375] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 974.920375] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 974.920375] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 974.920375] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 974.920375] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 974.921082] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: ddeaa346-8d42-43ad-b4d5-d70bc6618c67] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 974.948255] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 974.948489] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 974.948664] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-3bdd2488-a4c1-4666-8d60-7bd2f516269b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 976.858088] nova-conductor[52523]: Traceback (most recent call last): [ 976.858088] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 976.858088] nova-conductor[52523]: return func(*args, **kwargs) [ 976.858088] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 976.858088] nova-conductor[52523]: selections = self._select_destinations( [ 976.858088] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 976.858088] nova-conductor[52523]: selections = self._schedule( [ 976.858088] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 976.858088] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 976.858088] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 976.858088] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 976.858088] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.858088] nova-conductor[52523]: ERROR nova.conductor.manager [ 976.867124] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 976.867293] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 976.867472] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 976.926795] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: cbfd3e02-87d8-4728-bbb6-daaee4fd9c71] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 976.927551] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 976.927763] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 976.927926] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 976.933807] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 976.933807] nova-conductor[52523]: Traceback (most recent call last): [ 976.933807] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 976.933807] nova-conductor[52523]: return func(*args, **kwargs) [ 976.933807] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 976.933807] nova-conductor[52523]: selections = self._select_destinations( [ 976.933807] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 976.933807] nova-conductor[52523]: selections = self._schedule( [ 976.933807] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 976.933807] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 976.933807] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 976.933807] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 976.933807] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 976.933807] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 976.934859] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-a6a389ca-3a7b-445c-a0fd-218b828f1504 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: cbfd3e02-87d8-4728-bbb6-daaee4fd9c71] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.508684] nova-conductor[52522]: Traceback (most recent call last): [ 978.508684] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 978.508684] nova-conductor[52522]: return func(*args, **kwargs) [ 978.508684] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 978.508684] nova-conductor[52522]: selections = self._select_destinations( [ 978.508684] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 978.508684] nova-conductor[52522]: selections = self._schedule( [ 978.508684] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 978.508684] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 978.508684] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 978.508684] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 978.508684] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.508684] nova-conductor[52522]: ERROR nova.conductor.manager [ 978.517287] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 978.517574] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 978.517823] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 978.567015] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] [instance: cd7a65b6-27e4-4810-903c-4e60c6dfe2c3] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 978.567217] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 978.567441] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 978.567608] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 978.571176] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 978.571176] nova-conductor[52522]: Traceback (most recent call last): [ 978.571176] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 978.571176] nova-conductor[52522]: return func(*args, **kwargs) [ 978.571176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 978.571176] nova-conductor[52522]: selections = self._select_destinations( [ 978.571176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 978.571176] nova-conductor[52522]: selections = self._schedule( [ 978.571176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 978.571176] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 978.571176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 978.571176] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 978.571176] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 978.571176] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.572191] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-cbfec3d1-62d1-434d-815c-32f70e8bf489 tempest-InstanceActionsV221TestJSON-1420142507 tempest-InstanceActionsV221TestJSON-1420142507-project-member] [instance: cd7a65b6-27e4-4810-903c-4e60c6dfe2c3] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.884303] nova-conductor[52523]: Traceback (most recent call last): [ 978.884303] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 978.884303] nova-conductor[52523]: return func(*args, **kwargs) [ 978.884303] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 978.884303] nova-conductor[52523]: selections = self._select_destinations( [ 978.884303] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 978.884303] nova-conductor[52523]: selections = self._schedule( [ 978.884303] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 978.884303] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 978.884303] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 978.884303] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 978.884303] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.884303] nova-conductor[52523]: ERROR nova.conductor.manager [ 978.894459] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 978.894647] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 978.894853] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 978.941155] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 694bf83b-9958-4c99-8474-e9f200738f9d] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 978.941952] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 978.942200] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 978.942380] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 978.945924] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 978.945924] nova-conductor[52523]: Traceback (most recent call last): [ 978.945924] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 978.945924] nova-conductor[52523]: return func(*args, **kwargs) [ 978.945924] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 978.945924] nova-conductor[52523]: selections = self._select_destinations( [ 978.945924] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 978.945924] nova-conductor[52523]: selections = self._schedule( [ 978.945924] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 978.945924] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 978.945924] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 978.945924] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 978.945924] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 978.945924] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 978.946430] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-621529d8-94a3-443f-9519-69c42f5c86ba tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 694bf83b-9958-4c99-8474-e9f200738f9d] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 981.227358] nova-conductor[52522]: Traceback (most recent call last): [ 981.227358] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 981.227358] nova-conductor[52522]: return func(*args, **kwargs) [ 981.227358] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 981.227358] nova-conductor[52522]: selections = self._select_destinations( [ 981.227358] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 981.227358] nova-conductor[52522]: selections = self._schedule( [ 981.227358] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 981.227358] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 981.227358] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 981.227358] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 981.227358] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.227358] nova-conductor[52522]: ERROR nova.conductor.manager [ 981.241222] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 981.241222] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 981.241222] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 981.296758] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: d323828d-46fa-4238-b690-bd44e4f3c6f0] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 981.297387] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 981.297856] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 981.297856] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 981.300864] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 981.300864] nova-conductor[52522]: Traceback (most recent call last): [ 981.300864] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 981.300864] nova-conductor[52522]: return func(*args, **kwargs) [ 981.300864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 981.300864] nova-conductor[52522]: selections = self._select_destinations( [ 981.300864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 981.300864] nova-conductor[52522]: selections = self._schedule( [ 981.300864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 981.300864] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 981.300864] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 981.300864] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 981.300864] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 981.300864] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 981.301444] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-2c0b2e56-e152-45da-9ae9-e1fd3d89cdd3 tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: d323828d-46fa-4238-b690-bd44e4f3c6f0] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 982.574706] nova-conductor[52523]: Traceback (most recent call last): [ 982.574706] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 982.574706] nova-conductor[52523]: return func(*args, **kwargs) [ 982.574706] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 982.574706] nova-conductor[52523]: selections = self._select_destinations( [ 982.574706] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 982.574706] nova-conductor[52523]: selections = self._schedule( [ 982.574706] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 982.574706] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 982.574706] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 982.574706] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 982.574706] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.574706] nova-conductor[52523]: ERROR nova.conductor.manager [ 982.583551] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 982.583840] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 982.584028] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 982.631129] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] [instance: 2df9c235-9984-417f-9223-bbf3f6298dfc] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 982.631865] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 982.632084] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 982.632255] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 982.635734] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 982.635734] nova-conductor[52523]: Traceback (most recent call last): [ 982.635734] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 982.635734] nova-conductor[52523]: return func(*args, **kwargs) [ 982.635734] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 982.635734] nova-conductor[52523]: selections = self._select_destinations( [ 982.635734] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 982.635734] nova-conductor[52523]: selections = self._schedule( [ 982.635734] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 982.635734] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 982.635734] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 982.635734] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 982.635734] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 982.635734] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 982.636107] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-d81ae5c3-aa01-4aa4-a1b8-94b764e017d7 tempest-ServerActionsTestJSON-1715221641 tempest-ServerActionsTestJSON-1715221641-project-member] [instance: 2df9c235-9984-417f-9223-bbf3f6298dfc] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 983.599306] nova-conductor[52522]: Traceback (most recent call last): [ 983.599306] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 983.599306] nova-conductor[52522]: return func(*args, **kwargs) [ 983.599306] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 983.599306] nova-conductor[52522]: selections = self._select_destinations( [ 983.599306] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 983.599306] nova-conductor[52522]: selections = self._schedule( [ 983.599306] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 983.599306] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 983.599306] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 983.599306] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 983.599306] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.599306] nova-conductor[52522]: ERROR nova.conductor.manager [ 983.607851] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 983.607851] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 983.607851] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 983.649615] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 776c260f-aada-47b0-b8a8-db39ccf34aab] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 983.651533] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 983.651533] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 983.651533] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 983.654147] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 983.654147] nova-conductor[52522]: Traceback (most recent call last): [ 983.654147] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 983.654147] nova-conductor[52522]: return func(*args, **kwargs) [ 983.654147] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 983.654147] nova-conductor[52522]: selections = self._select_destinations( [ 983.654147] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 983.654147] nova-conductor[52522]: selections = self._schedule( [ 983.654147] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 983.654147] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 983.654147] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 983.654147] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 983.654147] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 983.654147] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 983.655221] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-9d41ceda-1c04-4b64-b019-01f9888d494b tempest-DeleteServersTestJSON-16586255 tempest-DeleteServersTestJSON-16586255-project-member] [instance: 776c260f-aada-47b0-b8a8-db39ccf34aab] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 989.644867] nova-conductor[52523]: Traceback (most recent call last): [ 989.644867] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 989.644867] nova-conductor[52523]: return func(*args, **kwargs) [ 989.644867] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 989.644867] nova-conductor[52523]: selections = self._select_destinations( [ 989.644867] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 989.644867] nova-conductor[52523]: selections = self._schedule( [ 989.644867] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 989.644867] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 989.644867] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 989.644867] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 989.644867] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.644867] nova-conductor[52523]: ERROR nova.conductor.manager [ 989.651818] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 989.652071] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 989.652250] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 989.734706] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] [instance: 73aa2208-f83a-440d-bcb1-b9ce2955cbe9] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 989.734706] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 989.734925] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 989.736704] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 989.742732] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 989.742732] nova-conductor[52523]: Traceback (most recent call last): [ 989.742732] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 989.742732] nova-conductor[52523]: return func(*args, **kwargs) [ 989.742732] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 989.742732] nova-conductor[52523]: selections = self._select_destinations( [ 989.742732] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 989.742732] nova-conductor[52523]: selections = self._schedule( [ 989.742732] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 989.742732] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 989.742732] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 989.742732] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 989.742732] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 989.742732] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 989.742732] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-e6df4a04-9048-4493-9456-b98eabedfcc1 tempest-AttachInterfacesUnderV243Test-54291424 tempest-AttachInterfacesUnderV243Test-54291424-project-member] [instance: 73aa2208-f83a-440d-bcb1-b9ce2955cbe9] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 991.037285] nova-conductor[52522]: Traceback (most recent call last): [ 991.037285] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 991.037285] nova-conductor[52522]: return func(*args, **kwargs) [ 991.037285] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 991.037285] nova-conductor[52522]: selections = self._select_destinations( [ 991.037285] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 991.037285] nova-conductor[52522]: selections = self._schedule( [ 991.037285] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 991.037285] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 991.037285] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 991.037285] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 991.037285] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.037285] nova-conductor[52522]: ERROR nova.conductor.manager [ 991.047689] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 991.047689] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 991.047689] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 991.104412] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] [instance: ec337a6a-1298-4ec4-8c12-a2ff363cab57] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 991.105141] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 991.105730] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 991.105730] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 991.109467] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 991.109467] nova-conductor[52522]: Traceback (most recent call last): [ 991.109467] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 991.109467] nova-conductor[52522]: return func(*args, **kwargs) [ 991.109467] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 991.109467] nova-conductor[52522]: selections = self._select_destinations( [ 991.109467] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 991.109467] nova-conductor[52522]: selections = self._schedule( [ 991.109467] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 991.109467] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 991.109467] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 991.109467] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 991.109467] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 991.109467] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 991.110017] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-e7caffd4-2cd2-4499-82bd-b6496f19ae6d tempest-ServersTestManualDisk-484959815 tempest-ServersTestManualDisk-484959815-project-member] [instance: ec337a6a-1298-4ec4-8c12-a2ff363cab57] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 992.424165] nova-conductor[52523]: Traceback (most recent call last): [ 992.424165] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 992.424165] nova-conductor[52523]: return func(*args, **kwargs) [ 992.424165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 992.424165] nova-conductor[52523]: selections = self._select_destinations( [ 992.424165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 992.424165] nova-conductor[52523]: selections = self._schedule( [ 992.424165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 992.424165] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 992.424165] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 992.424165] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 992.424165] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager result = self.transport._send( [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager raise result [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager Traceback (most recent call last): [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._select_destinations( [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager selections = self._schedule( [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.424165] nova-conductor[52523]: ERROR nova.conductor.manager [ 992.432123] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 992.432446] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 992.432670] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 992.487726] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] [instance: 625d0a76-795e-4f66-b9af-ce68f6945ab2] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52523) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 992.488400] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 992.488676] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 992.488895] nova-conductor[52523]: DEBUG oslo_concurrency.lockutils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52523) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 992.492155] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 992.492155] nova-conductor[52523]: Traceback (most recent call last): [ 992.492155] nova-conductor[52523]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 992.492155] nova-conductor[52523]: return func(*args, **kwargs) [ 992.492155] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 992.492155] nova-conductor[52523]: selections = self._select_destinations( [ 992.492155] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 992.492155] nova-conductor[52523]: selections = self._schedule( [ 992.492155] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 992.492155] nova-conductor[52523]: self._ensure_sufficient_hosts( [ 992.492155] nova-conductor[52523]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 992.492155] nova-conductor[52523]: raise exception.NoValidHost(reason=reason) [ 992.492155] nova-conductor[52523]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 992.492155] nova-conductor[52523]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 992.492928] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-0e664284-a49c-4023-9cda-7287b80f16c1 tempest-ServerAddressesNegativeTestJSON-630624972 tempest-ServerAddressesNegativeTestJSON-630624972-project-member] [instance: 625d0a76-795e-4f66-b9af-ce68f6945ab2] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 996.832021] nova-conductor[52522]: ERROR nova.scheduler.utils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 37916d26-1b5e-4991-83a2-ca5a5b00c2ac] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 37916d26-1b5e-4991-83a2-ca5a5b00c2ac was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 996.832021] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Rescheduling: True {{(pid=52522) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 996.832021] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 37916d26-1b5e-4991-83a2-ca5a5b00c2ac.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 37916d26-1b5e-4991-83a2-ca5a5b00c2ac. [ 996.832021] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 37916d26-1b5e-4991-83a2-ca5a5b00c2ac] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 37916d26-1b5e-4991-83a2-ca5a5b00c2ac. [ 996.868166] nova-conductor[52522]: DEBUG nova.network.neutron [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 37916d26-1b5e-4991-83a2-ca5a5b00c2ac] deallocate_for_instance() {{(pid=52522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 996.890163] nova-conductor[52522]: DEBUG nova.network.neutron [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 37916d26-1b5e-4991-83a2-ca5a5b00c2ac] Instance cache missing network info. {{(pid=52522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 996.893265] nova-conductor[52522]: DEBUG nova.network.neutron [None req-e59749df-cdb7-431a-98b4-6be65789904d tempest-ServersTestMultiNic-1605036069 tempest-ServersTestMultiNic-1605036069-project-member] [instance: 37916d26-1b5e-4991-83a2-ca5a5b00c2ac] Updating instance_info_cache with network_info: [] {{(pid=52522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.018042] nova-conductor[52522]: Traceback (most recent call last): [ 1000.018042] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.018042] nova-conductor[52522]: return func(*args, **kwargs) [ 1000.018042] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.018042] nova-conductor[52522]: selections = self._select_destinations( [ 1000.018042] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.018042] nova-conductor[52522]: selections = self._schedule( [ 1000.018042] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.018042] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1000.018042] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.018042] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1000.018042] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.018042] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.035404] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.035641] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.035826] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.082742] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] [instance: 6cc5b4fe-9a3a-468c-8a83-a758bd5b116d] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 1000.083478] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.083698] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.083897] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.087675] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 1000.087675] nova-conductor[52522]: Traceback (most recent call last): [ 1000.087675] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.087675] nova-conductor[52522]: return func(*args, **kwargs) [ 1000.087675] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.087675] nova-conductor[52522]: selections = self._select_destinations( [ 1000.087675] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.087675] nova-conductor[52522]: selections = self._schedule( [ 1000.087675] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.087675] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1000.087675] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.087675] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1000.087675] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.087675] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.088738] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] [instance: 6cc5b4fe-9a3a-468c-8a83-a758bd5b116d] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.119874] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.120133] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.120309] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.163648] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] [instance: 664f336f-543e-437e-bc9c-c41aad211626] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 1000.164367] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.164622] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.164835] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.170220] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 1000.170220] nova-conductor[52522]: Traceback (most recent call last): [ 1000.170220] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.170220] nova-conductor[52522]: return func(*args, **kwargs) [ 1000.170220] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.170220] nova-conductor[52522]: selections = self._select_destinations( [ 1000.170220] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.170220] nova-conductor[52522]: selections = self._schedule( [ 1000.170220] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.170220] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1000.170220] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.170220] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1000.170220] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.170220] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.170811] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] [instance: 664f336f-543e-437e-bc9c-c41aad211626] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.205448] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.205690] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.205864] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.263470] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] [instance: 51b9edcb-6ccf-41f5-b74b-574ffb4d7251] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 1000.264087] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.264314] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.264490] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.268589] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 1000.268589] nova-conductor[52522]: Traceback (most recent call last): [ 1000.268589] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.268589] nova-conductor[52522]: return func(*args, **kwargs) [ 1000.268589] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.268589] nova-conductor[52522]: selections = self._select_destinations( [ 1000.268589] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.268589] nova-conductor[52522]: selections = self._schedule( [ 1000.268589] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.268589] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1000.268589] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.268589] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1000.268589] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.268589] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.270279] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-c328d2ff-9ac9-4579-81e7-48503387acb6 tempest-ListServersNegativeTestJSON-1016907250 tempest-ListServersNegativeTestJSON-1016907250-project-member] [instance: 51b9edcb-6ccf-41f5-b74b-574ffb4d7251] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.918465] nova-conductor[52522]: Traceback (most recent call last): [ 1000.918465] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.918465] nova-conductor[52522]: return func(*args, **kwargs) [ 1000.918465] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.918465] nova-conductor[52522]: selections = self._select_destinations( [ 1000.918465] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.918465] nova-conductor[52522]: selections = self._schedule( [ 1000.918465] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.918465] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1000.918465] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.918465] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1000.918465] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.918465] nova-conductor[52522]: ERROR nova.conductor.manager [ 1000.925942] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.926314] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.926550] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.985593] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] [instance: 8453d04b-642a-48a9-a386-2633fa2483f0] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 1000.989017] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1000.989017] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1000.989017] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1000.991750] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 1000.991750] nova-conductor[52522]: Traceback (most recent call last): [ 1000.991750] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1000.991750] nova-conductor[52522]: return func(*args, **kwargs) [ 1000.991750] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1000.991750] nova-conductor[52522]: selections = self._select_destinations( [ 1000.991750] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1000.991750] nova-conductor[52522]: selections = self._schedule( [ 1000.991750] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1000.991750] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1000.991750] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1000.991750] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1000.991750] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1000.991750] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1000.992341] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-9e19c63e-f31d-49ac-9bde-ba8fb97762b6 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] [instance: 8453d04b-642a-48a9-a386-2633fa2483f0] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Failed to schedule instances: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1004.099176] nova-conductor[52522]: Traceback (most recent call last): [ 1004.099176] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1004.099176] nova-conductor[52522]: return func(*args, **kwargs) [ 1004.099176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1004.099176] nova-conductor[52522]: selections = self._select_destinations( [ 1004.099176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1004.099176] nova-conductor[52522]: selections = self._schedule( [ 1004.099176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1004.099176] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1004.099176] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1004.099176] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1004.099176] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 1654, in schedule_and_build_instances [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self._schedule_instances(context, request_specs[0], [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/conductor/manager.py", line 942, in _schedule_instances [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager host_lists = self.query_client.select_destinations( [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/client/query.py", line 41, in select_destinations [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager return self.scheduler_rpcapi.select_destinations(context, spec_obj, [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/rpcapi.py", line 160, in select_destinations [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager return cctxt.call(ctxt, 'select_destinations', **msg_args) [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/client.py", line 190, in call [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager result = self.transport._send( [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/transport.py", line 123, in _send [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager return self._driver.send(target, ctxt, message, [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 689, in send [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager return self._send(target, ctxt, message, wait_for_reply, timeout, [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 681, in _send [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager raise result [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager Traceback (most recent call last): [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager return func(*args, **kwargs) [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._select_destinations( [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager selections = self._schedule( [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager self._ensure_sufficient_hosts( [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager raise exception.NoValidHost(reason=reason) [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.099176] nova-conductor[52522]: ERROR nova.conductor.manager [ 1004.105839] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1004.106084] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1004.107528] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1004.146302] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] [instance: 8db5327f-4e7c-4686-adfb-ae86994e757e] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 1004.147044] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1004.147276] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1004.147448] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1004.152588] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] Failed to compute_task_build_instances: No valid host was found. There are not enough hosts available. [ 1004.152588] nova-conductor[52522]: Traceback (most recent call last): [ 1004.152588] nova-conductor[52522]: File "/usr/local/lib/python3.10/dist-packages/oslo_messaging/rpc/server.py", line 244, in inner [ 1004.152588] nova-conductor[52522]: return func(*args, **kwargs) [ 1004.152588] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 224, in select_destinations [ 1004.152588] nova-conductor[52522]: selections = self._select_destinations( [ 1004.152588] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 251, in _select_destinations [ 1004.152588] nova-conductor[52522]: selections = self._schedule( [ 1004.152588] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 452, in _schedule [ 1004.152588] nova-conductor[52522]: self._ensure_sufficient_hosts( [ 1004.152588] nova-conductor[52522]: File "/opt/stack/nova/nova/scheduler/manager.py", line 499, in _ensure_sufficient_hosts [ 1004.152588] nova-conductor[52522]: raise exception.NoValidHost(reason=reason) [ 1004.152588] nova-conductor[52522]: nova.exception.NoValidHost: No valid host was found. There are not enough hosts available. [ 1004.152588] nova-conductor[52522]: : nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1004.153141] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-bb942167-fbe3-4404-a357-ff83e3295359 tempest-AttachVolumeTestJSON-96048847 tempest-AttachVolumeTestJSON-96048847-project-member] [instance: 8db5327f-4e7c-4686-adfb-ae86994e757e] Setting instance to ERROR state.: nova.exception_Remote.NoValidHost_Remote: No valid host was found. There are not enough hosts available. [ 1008.802712] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Took 0.12 seconds to select destinations for 1 instance(s). {{(pid=52522) _schedule_instances /opt/stack/nova/nova/conductor/manager.py:945}} [ 1008.838586] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1008.838816] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1008.838988] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1008.862035] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1008.862294] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1008.862485] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1008.862839] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1008.863039] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1008.863206] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1008.871839] nova-conductor[52522]: DEBUG nova.quota [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Getting quotas for project 1ebb28978bff4e9a93497373d9a16e18. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:393}} [ 1008.874225] nova-conductor[52522]: DEBUG nova.quota [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Getting quotas for user 023b6b895086429595d77b626450d972 and project 1ebb28978bff4e9a93497373d9a16e18. Resources: {'cores', 'instances', 'ram'} {{(pid=52522) _get_quotas /opt/stack/nova/nova/quota.py:383}} [ 1008.880066] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] [instance: 8b64034a-4d67-4605-adb8-a007dd735230] Selected host: cpu-1; Selected node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28; Alternates: [] {{(pid=52522) schedule_and_build_instances /opt/stack/nova/nova/conductor/manager.py:1761}} [ 1008.880518] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1008.880726] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1008.880898] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1008.883497] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] [instance: 8b64034a-4d67-4605-adb8-a007dd735230] block_device_mapping [BlockDeviceMapping(attachment_id=,boot_index=0,connection_info=None,created_at=,delete_on_termination=True,deleted=,deleted_at=,destination_type='local',device_name=None,device_type='disk',disk_bus=None,encrypted=False,encryption_format=None,encryption_options=None,encryption_secret_uuid=None,guest_format=None,id=,image_id='ce78d8ba-df84-4ce9-9b5e-632fda86b4cc',instance=,instance_uuid=,no_device=False,snapshot_id=None,source_type='image',tag=None,updated_at=,uuid=,volume_id=None,volume_size=None,volume_type=None)] {{(pid=52522) _create_block_device_mapping /opt/stack/nova/nova/conductor/manager.py:1506}} [ 1008.884180] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1008.884385] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1008.884554] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1008.896319] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Acquiring lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:404}} [ 1008.896522] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:409}} [ 1008.896693] nova-conductor[52522]: DEBUG oslo_concurrency.lockutils [None req-42128840-7836-452c-b26a-f75d3184a2f3 tempest-AttachInterfacesV270Test-252724916 tempest-AttachInterfacesV270Test-252724916-project-member] Lock "74c00a1b-a396-4090-b2c1-a2ad7f5ace70" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.000s {{(pid=52522) inner /usr/local/lib/python3.10/dist-packages/oslo_concurrency/lockutils.py:423}} [ 1043.180413] nova-conductor[52522]: ERROR nova.scheduler.utils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: 3ab1b905-cd6f-4d2b-a244-f85e56f796d3] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 3ab1b905-cd6f-4d2b-a244-f85e56f796d3 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 1043.181244] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Rescheduling: True {{(pid=52522) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 1043.181421] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 3ab1b905-cd6f-4d2b-a244-f85e56f796d3.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 3ab1b905-cd6f-4d2b-a244-f85e56f796d3. [ 1043.183414] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: 3ab1b905-cd6f-4d2b-a244-f85e56f796d3] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 3ab1b905-cd6f-4d2b-a244-f85e56f796d3. [ 1043.208910] nova-conductor[52522]: DEBUG nova.network.neutron [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: 3ab1b905-cd6f-4d2b-a244-f85e56f796d3] deallocate_for_instance() {{(pid=52522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1043.226616] nova-conductor[52522]: DEBUG nova.network.neutron [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: 3ab1b905-cd6f-4d2b-a244-f85e56f796d3] Instance cache missing network info. {{(pid=52522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1043.229921] nova-conductor[52522]: DEBUG nova.network.neutron [None req-6f634d99-9457-48c6-a6d2-9ded8cd6c6c3 tempest-ServersAdminTestJSON-1213463885 tempest-ServersAdminTestJSON-1213463885-project-member] [instance: 3ab1b905-cd6f-4d2b-a244-f85e56f796d3] Updating instance_info_cache with network_info: [] {{(pid=52522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.148650] nova-conductor[52523]: ERROR nova.scheduler.utils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] [instance: 036d6de2-f69b-4714-b89e-9c4307253675] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 036d6de2-f69b-4714-b89e-9c4307253675 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 1094.149228] nova-conductor[52523]: DEBUG nova.conductor.manager [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Rescheduling: True {{(pid=52523) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 1094.149455] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 036d6de2-f69b-4714-b89e-9c4307253675.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 036d6de2-f69b-4714-b89e-9c4307253675. [ 1094.149685] nova-conductor[52523]: WARNING nova.scheduler.utils [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] [instance: 036d6de2-f69b-4714-b89e-9c4307253675] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 036d6de2-f69b-4714-b89e-9c4307253675. [ 1094.174705] nova-conductor[52523]: DEBUG nova.network.neutron [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] [instance: 036d6de2-f69b-4714-b89e-9c4307253675] deallocate_for_instance() {{(pid=52523) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1094.307242] nova-conductor[52523]: DEBUG nova.network.neutron [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] [instance: 036d6de2-f69b-4714-b89e-9c4307253675] Instance cache missing network info. {{(pid=52523) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1094.311250] nova-conductor[52523]: DEBUG nova.network.neutron [None req-f8a499f7-7ae1-4eaf-9aab-36371b84c6c5 tempest-VolumesAssistedSnapshotsTest-2138134395 tempest-VolumesAssistedSnapshotsTest-2138134395-project-member] [instance: 036d6de2-f69b-4714-b89e-9c4307253675] Updating instance_info_cache with network_info: [] {{(pid=52523) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.714014] nova-conductor[52523]: DEBUG nova.db.main.api [None req-0ddcd069-9e43-4f7b-afad-2168625b0fa4 tempest-ServersTestBootFromVolume-728161428 tempest-ServersTestBootFromVolume-728161428-project-member] Created instance_extra for e259637a-0fc8-4368-8a7a-c15a134ed17d {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1142.452900] nova-conductor[52522]: ERROR nova.scheduler.utils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] [instance: 5888cc9f-7341-4f9c-a93c-dd5ec95f7369] Error from last host: cpu-1 (node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28): ['Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2611, in _build_and_run_instance\n self.driver.spawn(context, instance, image_meta,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 529, in spawn\n self._vmops.spawn(context, instance, image_meta, injected_files,\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 786, in spawn\n self._fetch_image_if_missing(context, vi)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 639, in _fetch_image_if_missing\n image_cache(vi, tmp_image_ds_loc)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 537, in _cache_sparse_image\n vm_util.copy_virtual_disk(\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1423, in copy_virtual_disk\n session._wait_for_task(vmdk_copy_task)\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task\n return self.wait_for_task(task_ref)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 398, in wait_for_task\n return evt.wait()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/event.py", line 125, in wait\n result = hub.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/eventlet/hubs/hub.py", line 313, in switch\n return self.greenlet.switch()\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner\n self.f(*self.args, **self.kw)\n', ' File "/usr/local/lib/python3.10/dist-packages/oslo_vmware/api.py", line 448, in _poll_task\n raise exceptions.translate_fault(task_info.error)\n', "oslo_vmware.exceptions.VimFaultException: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2426, in _do_build_and_run_instance\n self._build_and_run_instance(context, instance, image,\n', ' File "/opt/stack/nova/nova/compute/manager.py", line 2718, in _build_and_run_instance\n raise exception.RescheduledException(\n', "nova.exception.RescheduledException: Build of instance 5888cc9f-7341-4f9c-a93c-dd5ec95f7369 was re-scheduled: A specified parameter was not correct: fileType\nFaults: ['InvalidArgument']\n"] [ 1142.453620] nova-conductor[52522]: DEBUG nova.conductor.manager [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Rescheduling: True {{(pid=52522) build_instances /opt/stack/nova/nova/conductor/manager.py:695}} [ 1142.454036] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] Failed to compute_task_build_instances: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 5888cc9f-7341-4f9c-a93c-dd5ec95f7369.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 5888cc9f-7341-4f9c-a93c-dd5ec95f7369. [ 1142.454314] nova-conductor[52522]: WARNING nova.scheduler.utils [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] [instance: 5888cc9f-7341-4f9c-a93c-dd5ec95f7369] Setting instance to ERROR state.: nova.exception.MaxRetriesExceeded: Exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance 5888cc9f-7341-4f9c-a93c-dd5ec95f7369. [ 1142.474698] nova-conductor[52522]: DEBUG nova.network.neutron [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] [instance: 5888cc9f-7341-4f9c-a93c-dd5ec95f7369] deallocate_for_instance() {{(pid=52522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1142.528221] nova-conductor[52522]: DEBUG nova.network.neutron [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] [instance: 5888cc9f-7341-4f9c-a93c-dd5ec95f7369] Instance cache missing network info. {{(pid=52522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1142.531484] nova-conductor[52522]: DEBUG nova.network.neutron [None req-68696909-aa75-4998-bb13-0c63cf192b9e tempest-TenantUsagesTestJSON-787131523 tempest-TenantUsagesTestJSON-787131523-project-member] [instance: 5888cc9f-7341-4f9c-a93c-dd5ec95f7369] Updating instance_info_cache with network_info: [] {{(pid=52522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.646912] nova-conductor[52523]: DEBUG nova.db.main.api [None req-30b24ba5-531e-443c-a69b-a353e14a01d7 tempest-ServerAddressesTestJSON-1201308295 tempest-ServerAddressesTestJSON-1201308295-project-member] Created instance_extra for 08e2d758-9005-4822-b157-84710b9c5ed4 {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1142.960739] nova-conductor[52523]: DEBUG nova.db.main.api [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Created instance_extra for 54919bf0-b9f3-4bfc-ba1a-c6a52013e351 {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1143.208139] nova-conductor[52523]: DEBUG nova.db.main.api [None req-486ce0a8-156c-48a6-abf8-e5e7d3003872 tempest-MultipleCreateTestJSON-328947127 tempest-MultipleCreateTestJSON-328947127-project-member] Created instance_extra for a925d5fc-6437-40bb-adf1-ea10c32dde2a {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1192.755710] nova-conductor[52522]: DEBUG nova.db.main.api [None req-2fffb6e4-07b3-4aeb-a603-d93f7172df13 tempest-ServersNegativeTestMultiTenantJSON-1610443141 tempest-ServersNegativeTestMultiTenantJSON-1610443141-project-member] Created instance_extra for 076c3dd5-9043-456d-af24-0d2273321085 {{(pid=52522) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1192.991993] nova-conductor[52523]: DEBUG nova.db.main.api [None req-b6b9848f-f046-48a5-90f1-4a5a47dde445 tempest-ServersTestFqdnHostnames-309226866 tempest-ServersTestFqdnHostnames-309226866-project-member] Created instance_extra for fcf47169-eb7a-4644-bf3f-7150c44c247f {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1193.793834] nova-conductor[52523]: DEBUG nova.db.main.api [None req-645fc93b-ea81-455c-ada2-ce9f7feb6a40 tempest-DeleteServersAdminTestJSON-4043494 tempest-DeleteServersAdminTestJSON-4043494-project-member] Created instance_extra for d55ee9a1-6921-4648-ace2-f2da13c3523e {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1194.046707] nova-conductor[52523]: DEBUG nova.db.main.api [None req-5bf779b6-749c-4443-bf58-24963e6d36c6 tempest-ServerPasswordTestJSON-1793502215 tempest-ServerPasswordTestJSON-1793502215-project-member] Created instance_extra for e8ed8e30-1b1d-4ab4-abd7-e68cb72916cb {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}} [ 1194.292132] nova-conductor[52523]: DEBUG nova.db.main.api [None req-683d5b43-8eb8-4924-b841-8951f706b180 tempest-AttachVolumeNegativeTest-692012757 tempest-AttachVolumeNegativeTest-692012757-project-member] Created instance_extra for 54ded864-1c3e-4a47-968f-ca597c82cb87 {{(pid=52523) instance_extra_update_by_uuid /opt/stack/nova/nova/db/main/api.py:2551}}